26 Comments
Oct 5, 2023·edited Oct 5, 2023Liked by Gary Marcus

Excellent piece. I have come to understand that all the great results attributed to LLMs should instead be attributed to the millions of human beings that LLMs use as preprocessors. LLMs are essentially cheaters with no understanding. What LLMs do show is that human language is highly statistical. But this is not a great scientific breakthrough since linguists have known this for decades, if not centuries.

The hype surrounding LLMs (and generative AI in general) always brings me to what I consider to be the acid test for AGI. Can it be used to design a robot cook that can walk into an unfamiliar kitchen and fix a meal? Generative AI models don't stand a chance to pass this test. Not in a billion years.

Expand full comment

To say that a language model has a model of the world is an oxymoron. There is a general principle here. The general principal is that all of these claims that some GenAI model has a cognitive property are all based on the logical fallacy of affirming the consequent. Here is an example of this pseudologic. If Lincoln was killed by robots, then Lincoln is dead. Lincoln is dead, therefore, he was killed by robots. This conclusion is obviously nonsense.

In the case of GenAI, the argument is: If the model has this cognitive property (reasoning, sentience, model of the world, etc.), then it will answer this question correctly. It answers correctly, therefore it has the described competence. This conclusion is no more valid than the one about Lincoln's death. It is not valid to assume from the consequent, that the premise is true. Other factors could have produced the result (John Wilkes Booth or language patterns).

To assert that large language models have any properties beyond those that were designed into them (language modeling) is magical thinking based on a logical fallacy. The language model is sufficient to explain the observation, and since we know how the language models was built, we have no reason to think that there is any more to it than that.

Expand full comment

I don't always agree with your positions, but I think you are doing a valuable service in providing a narraticve counter to the hype train that runs around papers like this. Keep up the good work!

Expand full comment

So, as I was reading your article, Gary, it triggered a memory for me. Before I learned how to drive, the map I had in my head about the world was pretty limited. I knew my way to and from school on my bike, but I didn't really understand where other cities were. I would hear my parents and their friends talk about these places, but I had no firm grasp on them. I didn't have a model/map in my head, in other words, like I did for the bike route to/from school. I knew that my parents would say that such and such a town was to the north/south/east/west of my hometown, but I had no idea which roads to take to get there or how far away they might be beyond my parents estimated driving time. In other words, I had learned what they had said about these towns, but I was still lacking a detailed model with which to reason about reality. Any model I had was based on language and was extremely "low resolution."

Expand full comment
Oct 5, 2023Liked by Gary Marcus

It is interesting that GPT-4 understands the difference between, say, 1a. "Amy poured water into the glass" (content locative) and 2a. "Amy filled the glass with water" (container locative), in that in 2a., Amy fills the whole glass. When you ask it whether 1b. *"Amy poured the glass with water" and 2b. *"Amy filled water into the glass", it says that they are both grammatical (but that they might sound odd, and it gives some explanations as to why, that are not convincing). Even young children, who might never have heard these sentences, understand that both 1b. and 2b. are wrong, based on the causality of language and their model of the world. That's a very strong argument for there being some kind of grammatical generative system in our minds that GPT-4 lacks, and no matter how much data we fill it with, it's not going to get it.

Expand full comment
Oct 5, 2023Liked by Gary Marcus

I'm reminded of a remark the Bill Powers made to me years ago in connection with an Old School symbolic model I had constructed for the purpose of analyzing the semantics of a Shakespeare sonnet. Powers remarked (http://www.jstor.org/stable/2907155):

"There are always two levels of modelling going on. At one level, modelling consists of constructing a structure that, by its own rules, would behave like the system being modelled, and if one is lucky produce that behavior by the same means (the same inner processes) as the system being modelled. That kind of model "runs by itself"; given initial conditions,the rules of the model will generate behavior.

"But the other kind of modelling is always done at the same time: the modeller provides for himself some symbolic scheme together with rules for manipulating the symbols,f or the purpose of reasoning about the other kind of model. The relationship between the two kinds of models is very much like the relationship you describe between the thought-level and the abstraction-level.

"The biggest problem in modelling is to remain aware of which model one is dealing with. Am I inside my own head reasoning about the model, or am I inside the model applying its rules to its experiences? This is especially difficult to keep straight when one is talking about cognitive processes; unless one is vividly aware of the problem one can shift back and forth between the two modes of modelling without realizing it."

In this case, as Davis remarks, it is not at all obvious that the LLM itself has explicit access to this world "model" that is so obvious to an external observer, an observer standing in a "transcendent" relationship to the model. I fear that this kind of confusion is very common in thinking about LLMs and is responsible for over-estimating their capacities.

Expand full comment

Love it. Stray parenthesis here:

Finding that some stuff (correlates

Expand full comment

Originally, OpenAI tried to get to safety in GPT3 using fine-tuning only. That was so 'jailbreaking-prone' that they had to install some simplistic filtering (like filtering on 'bad words'). I call these filters 'AoD-filters', where 'AoD' stands for 'Admission of Defeat'. Most illustrative here is that they not only filter the prompt, but they also filter the generation that way. Hence GPT creates a reply which is then flagged by GPT's *own* 'AoD' filter. Funny and telling.

It is difficult for people not to get bewitched by these systems. I recall your piece about 'How not to test GPT" post a while back.

LLMs are 'statistically constrained hallucinators'. The constraining will realistically never scale to the amount that it becomes a real model that has logical understanding. Even in OpenAI's 'LLMs are few shot learners' article, you are easily misled by the fact that the X-axis is logarithmic... If you fix that, interesting visuals appear...

Expand full comment

Indeed, all LLM does here is find spatial correlations between words used together. That is notable, but like all other emerging skills, is not altogether surprising.

LLM lacks spatial reasoning skills, as it only models language. Granted, art generation programs show that spatial logic can be learned, but not with a model that only uses words.

Expand full comment

Oh, man. Thanks for putting in the work to clarify this. I saw the post, and my first thought was confusing word relations with an actual model or formal intrinsic representation. There is a lot to be said regarding how a physics or a spatial model works, how an artificial neural network represents a model, and the distance to how the human brain organizes these ideas. Part of the current hype forced exciting ideas forward but lacking real substance.

Expand full comment

OthelloGPT would like a word...

See https://thegradient.pub/othello/

Expand full comment

Also this week, the advent of Assembly Theory was published in the journal *Nature*, unifying biology with physics. The treatment of LLMs as singular entities is unrealistic. Now and forever they will be collaborators, as are all of us. Such is the essence of complexity. The pursuit of singular infallible causal models is equally illusory, having been diligently pursued since at least the time of Aristotle without success. Progress will proceed through growth and refinement: With reinforcement learning from human feedback (RLHF), constitutional and multiagent frameworks, etc.

Expand full comment

Using things like clustering in the 70's to find Dallas and Austin are close from texte.

Expand full comment

This kind of paper, as the one written by Gurnee & Tegmark and discussed above, besides its technical content, expresses a form of enthusiasm about the future applications of AI and the propensity to see mainly the positive aspects of AI. As I wrote in a preceding comment, the AI-driven tools will be soon so easy to handle, so helpful, and so apparently efficient, that a lot of persons will gladly use them even some very serious issues remain, also in a professional context. While AI-based tools will be more and more extensively used, there will be less and less space for criticism. People will not like to hear criticism about handy tools they use every day and are satisfied with. We can only hope, and not be certain, that the AI systems designers will incorporate some logical inference scheme in the LLMs.

Using the physical science terminology, the LLMs are kind of empirical models as opposed to knowledge based (physical laws) models. The empirical models are correlations based on experiments’ results. They were sometimes called “black box models”. They are a set of polynomial equations with a multitude of coefficients obtained by mathematical fitting the model output to some experimental data. They can be efficient for prediction if the user does not cross (even by the smallest amount) the value domain of parameters considered and does not try to represent a situation where an additional parameter is needed (even just a single parameter more). We have the same limitations in LLMs. In physics, black box models are used to describe a single phenomenon, a single process. But the ambition with LLMs is quite extravagant as they will pretend to describe the entire world. If one wants to describe the entire world with a “black box model”, one would need an infinite number of parameters and an infinite number of coefficients. That’s why obviously we will need a hybrid approach to get an efficient and reliable AI system, a combination of empirical correlations, knowledge based equations and logical rules.

Expand full comment

A set of comparator functions modeled as a "brain" that's unattached to any somatic basis at all--much less one resembling a "nervous system"--receiving exclusively verbal/numerical information input to describe an external field that the comparator functions don't even recognize as an external field = ArtificiaI Intelligence

Don't get me wrong. Despite those inherent constraints, AI is potentially a valuable tool, given properly directed focus from an external human intelligence.

I'd especially like to see a concerted effort to drill an AI program in the principles of logical fallacy detection. The evaluations might prove to be vulnerable to occasional reasoning flaws at the outset, but I'm confident of the ability of an AI program to learn and improve in its ability at logical fallacy detection (although it's likely to demand more clarity in the semantics of the input being screened.)

Significantly, the default state of AI harbors an advantage that individual human awarenesses find very difficult to consistently maintain: Impartiality.

The rules of logical fallacy detection are, for the main part, simple and straightforward.* The lists are overwhelmingly similar.

example https://www.logicalfallacies.org/

However, as far as the real world application of logical fallacy detection, the obstacle that is most liable to confound the intelligence of humans is the challenge of applying the rules consistently. Time and again I've read humans engaged in debate (often of a political nature) doing takedowns of arguments by the side they oppose with unerring precision- but they're unable to critique their own positions with the same level of incisive reasoning and logical rigor. The limitations of ego-based human standpoint tend to get in the way. With diligent practice and self-discipline it's a fixable glitch, theoretically. But considered as a latent potential, idea-clutching personal egos always present a problem for us humans. A formidable obstacle to lucidity and appropriately responsible action.

By contrast, in the case of Artificial Intelligence, I don't feature an autonomous Ego-based agenda as presenting a problem at all, because there's no evidence that any AI program possesses an Ego. Hence, AI is potentially available as a tool to expose the flaws in any fact claim, inference, or debate position, without fear or favor.

[ * with some exceptions. For instance, not every "slippery slope" is a fallacy; some slippery slopes indicate actual perils. ]

Expand full comment