Is Generative AI Logical?

 

 

 

The French have a marvelous phrase, “Remise en question”, to call into question. At the beginning of the twentieth century, that phrase was indeed very much the fact. That fact produced great science but, of course, horrible politics. By assuming that electrons were not simple planetary spheres whirling around an atom core, but clouds of probability, science produced quantum physics, the computer revolution, and the generative AI of the 21st century. But what is a likely development of AI? The following discusses the philosophical foundations of generative AI and therefore its advantages and practical limitations.

    

 

Philosophical Foundations of Generative AI 

 

Ludwig Wittgenstein founded the science of linguistics. The Wittgensteins were part of an Austrian Habsburg civilization that produced such notables as Freud, Klimt, and Bruckner. According to Janik and Toulmin, “Ludwig Wittegenstein’s (grandly named) Tractatus Logico-Philosophicus…intimately and consciously related to…the critique of language and society…” 1 Then, as now, there was a lot of criticism about everything.

 

Wittgenstein set out to do for language what the German scientist, Heinrich Hertz, did in The Principles of Mechanics. He turned mechanics into a perfectly mathematical system involving the universal primitives of, “time, space, and mass as objects of experience.” 2  Wittgenstein’s effort was not successful; for in Philosophical Investigations he concluded that people flexibly use languages in different social ways that depend upon context, a word that is very often used in the 21st century and is fundamental to the transformer architecture of generative AI.

 

But as Bain and Festa of Oregon State note, “There is seemingly a gap between (AI’s) ability to model language and the desired behavior we would expect of such a model.” 3 What generative AI lacks is syntax, or a set of rules that determine how languages, and therefore words, are put together. Putting a thought into linguistic form, a sentence in any language must contain a subject and a verb, such as a implies b. The consequences of generative AI’s bottom-up vector driven approach are both a fundamental lack of top-down logic and a lack of innate morality.  As the question was broached in high school, what distinguishes a human from other animals? Here, it is language, syntax, logic and therefore communication.

 

Noam Chomsky is a famous linguist and professor emeritus of MIT. In a 3/8/23 NYT article titled, “The False Promise of ChatGPT”, he writes as one who believes in human nature and top-down universals. This is an important article, also for its consideration of scientific explanation. Professor Chomsky does not like generative AI.

 

1)     “The human mind is not, like ChatGPT and its ilk, a lumbering statistical engine for pattern matching, gorging on hundreds of terabytes of data and extrapolating the…most probable answer to a scientific question.” A model doesn’t simply get “smarter” because it has more data. To cite an example of statistical pattern matching, “Whereas humans are limited in the kinds of explanation we can rationally conjecture, machine learning systems can learn both that the earth is flat and that the earth is round. They trade merely in probabilities that change over time.”

 

2)    “On the contrary, the human mind is a surprisingly efficient and even elegant system that operates with small amounts of information; it seeks not to infer brute correlations among (vector) data points but to create explanations (What is going on here?)….When linguists seek to develop a theory for why a given language works as it does, they are building consciously and laboriously an explicit version of the grammar that the child builds instinctively and with minimal exposure to information. The child’s operating system is completely different from that of a machine learning program. Indeed, such programs are stuck in a prehuman or non human phase of cognitive evolution.” They don’t, like living animals (like Fido), know what goals to seek.

 

3)    The author offers this example about true science, which is about logical explanation. The science, merely of correlation, is a “pseudoscience” because, as we learned, “correlation is not causation.”

 

a)    I hold an apple in my hand.

 

b)    The apple will fall if I open my hand (that is correlation and prediction; that is as far as generative AI can go).

 

c)    The apple will fall if I open my hand because (of the force of gravity or the curvature of space-time). These are explanations (about the universe or ordinary life, and very elegant ones at that). Furthermore, each separate theory has mountains of empirical backup data. The universe may be complex; but the reason, as Hertz notes, why scientists continually seek fundamentals is to simplify the theory.

 

4)    Chatbots lack morality. This is why so many people have been grappling with the question of “controlling” chatbots. “Note, for all the seemingly sophisticated thought and language, the moral indifference born of unintelligence. Here, ChatGPT exhibits something like the banality of evil:  plagiarism and apathy and obviation. It summarizes the standard arguments in the literature by a kind of super-autocomplete, refuses to take a stand on anything (that really matters), pleads not merely ignorance but lack of intelligence and ultimately offers a ‘just following orders’ defense, shifting responsibility to its creators (where it belongs anyway).”

 

Ultimately, generative AI lacks logic. As we noted, it lacks syntax and therefore the constraints of logic. It also lacks the logic of simple arithmetic (+,-,x,¸). If this is true, then there are real problems in using chatbots in big decisions.

 

Apple Computer has published a paper that shows chatbots lack the ability to do simple arithmetic. Samy Bengio is the senior director of AI at Apple. Bengio et al note this in, “GSM-Symbolic: Understanding the Limitations of Mathematical Reasoning in Large Language Models”:

 

·      The GSM8K benchmark is widely used to assess the mathematical reasoning of models on grade-school-level questions….we introduce GSM-Symbolic, an improved benchmark created from symbolic templates that allow for the  generation of a diverse set of questions….Our findings reveal that LLMs exhibit notable variance when responding to different instantiations of the same question….we investigate the fragility of mathematical reasoning in these models and demonstrate that their performance significantly deteriorates as the number of clauses in a question increases. We hypothesize that this decline is due to the fact current LLMs are not capable of genuine logical reasoning; instead, they attempt to replicate the reasoning steps observed in their training data. When we add a single clause that appears relevant to the question, we observe significant performance drops (up to 65%) across all state-of-the-art models.

 

But the current demonstrated ability of the Perplexity chatbot to do arithmetic and calculus is probably due to a logical add-on, rather than to a fundamental change in the chatbot’s bottom-up and agnostic nature. The Perplexity “Pro-Search” feature uses top-down reasoning strategies to a point, when bottom-up AI then takes over.

 

Communicated scientific research is more clearly constrained by a statistical top-down concept: the null hypothesis, H0.

 

Advantages and Practical Limitations   

 

“Correlation is not causality.” In generative AI, correlation is probably enough to produce usable products (people also having habits). But generative AI, lacking logic, cannot be more superior than humans to pull out causality from a complex world, to also determine, using Mervyn King’s words, “What is going on here?”

 

The editor of Scientific American wrote, “…(what) I enjoy most…is learning how scientists do their work - the inspirations the questions, the insights, the collaborations, the, ‘hmm, that’s strange’ moments.” 4

 

Consumer Purposes

 

Our Use:

 

But the chatbot is a very useful focused information source. It is also useful to research extensions of the theories we learned in school. In both senses, derived from the data on the internet (and not other chatbots), it is like the Great Library of Alexandria in the ancient world of 2nd to 3rd century B.C., a universal library housing between 40,000-400,000 scrolls. If you can imagine yourself as a researcher entering the library. What you will find there is what you can make of it, depending on what questions you ask. The library, itself, doesn’t act as an oracle.

 

Consider the following exchange we had with the Perplexity chatbot on 1/4/25:

 

Us: Can you give me a language where to complete a thought, there is no verb? An one to one correspondence between an object and its name doesn’t count.

 

Perplexity: After citing sources: While there are no known natural languages that completely lack verbs, there is an artificial language called Kelen that was designed to function without traditional verbs…However its important to note that the classification of Kelen as truly verbless is debated among linguists…it’s worth noting that this structure is not found in natural human languages. All known natural languages use verbs in some form, suggesting that verbs play a crucial role in human communication and cognition.

 

We can imagine trying to research this question, unaided.

 

General Use:

 

·      Correlation is not the structure of causality. For social science purposes, a model correlation of less than .66 is not useful for theory. But people also having habits, pure correlation might be useful for agentic AI; because agents might be helpful for how people act most of the time, on the computer. It would be interesting to also ask how computer agents might be connected to life, where both the (hopefully) real facts you select and ethics matter.

·      Chatbots are currently used in academia to produce, from original thoughts and research of course, rough first drafts of an essay. They are also useful sources for new theories, materials and interesting relationships.

·      Census Bureau research on 11/28/23, regarding the economic adoption of generative AI indicates that only 3.8% of all businesses currently use the technology; but in information and technology, it is used by 13.8% of all businesses.5 On the other hand, a private company online commercial survey found that 39 percent of U.S. population age 18-64 used generative AI in 8/24, an adoption even faster than the adoption of the internet.6 These findings can all be true, but for different populations. Here, we consider the population of large (S&P 500) companies and of smaller companies.

 

Commercial Use in Large Companies

 

For large companies: Return on Investment >> Cost of Capital, the data must be in good shape (FAIR), and there must be a causal logic that justifies millions of dollars of expenditures.

 

A Goldman Sachs 8/5/24 study indicates that AI has, “…some cool applications…But there isn’t this killer application that’s consuming a lot of capacity right away….there’s a race to see who can build the best foundational model (general models that can be applied to many applications). That race isn’t going to slow down anytime soon. From an ROI perspective, if you look at it over the next one or two years, maybe the ROI isn’t great. But if you have a return stream of 20 years (Wall Street?) associated with your building the best tech stack today (costing hundreds of billions of dollars), then certainly you could justify the investment.”

 

We will comment on this later.

 

Smaller Companies

 

Text-based AI packages can provide management with a more detailed look, that is more easily implemented in a smaller company. For instance, agentic AI can help determine what a marketing force is doing in a sales territory.

 

Speed of Adoption

 

In both cases, AI adoption is likely to take time. In the larger companies, the overall company data must be in shape and implementation takes time. In the smaller companies, management may not see AI as beneficial.

 

Cost of Resources

 

According to the 12/20/24 WSJ, OpenAI’s latest model, GPT-5 requires a six month training run costing about 500 million dollars in computing costs alone. By increasing the data (which the models are running out of) or the parameters (neurons) in the model, it is hoped, “…it will make fewer mistakes…(and is supposed to get) smarter….Across the industry, a debate is raging over whether improvements in AIs is starting to plateau. Sutskever…declared at a recent AI conference that the age of maximum data is over…. ‘You can even go as far as to say that data is the fossil fuel of AI.’ And that fuel was starting to run out.”

 

Perhaps generative AI should be used to augment, rather than replace, human capabilities. It makes no sense to pour hundreds of billions of dollars into AI chips, data centers and electric grids in the hope of a “smarter” AI.

 

Climate Change

 

Coping with a changing world requires clear logic, a implies b, that is, C02  growth implies average temperature growth. To consider temperature growth “politically determined” is to ignore objective reality and is to make a very large mistake. We have mentioned a major philosophical problem with generative AI, a lack of syntax, that is likely to affect the scope of its applicability. From the standpoint of society, the excessive cash flow spent on AI would be better spent on clearing the brush in California, building sea walls in Miami and Manhattan.

 

Large fires in California, massive tornados in the Midwest, and floods in Miami. That is the future, if the climate becomes more extreme.

      

 

 

Footnotes

 

 

1. Allan Janik and Stephen Toulmin; “Wittgenstein’s Vienna”; Simon and Shuster; New York, NY; 1973; p. 93.

 

2. Heinrich Hertz; “The Principles of Mechanics”; London: Macmillan; 1899; p. 26, Introduction.

 

Professor Hertz, in the Introduction, also first states the essentials of “models” and “independent variables”, concepts that are commonplace today. He was noted for enabling radio transmissions by Maxwell’s electromagnetic waves. U.S. alternating current also oscillates at a rate of 60 (Hz) per second on the grid.

 

3. Robert Bain et al.; “Wittgenstein’s Philosophy of Language The Philosophical Origins of Modern NLP Thinking”; oregonstate.edu; p. 7.

 

4. Scientific American; editorial; January, 2025; p.4.

 

5.. Cory Breaux and Emin Dinlersoz; “How Many U.S. Businesses Use Artificial Intelligence?”; U.S. Government Census Bureau website; 11/28/23.

 

6. Alexander Bick et al.; “The Rapid Adoption of Generative AI”; Federal Reserve Bank of St. Louis; 9/20/24.