Comment by adamzwasserman
5 days ago
The article misses three critical points:
1. Conflates consciousness with "thinking" - LLMs may process information effectively without being conscious, but the article treats these as the same phenomenon
2. Ignores the cerebellum cases - We have documented cases of humans leading normal lives with little to no brain beyond a cerebellum, which contradicts simplistic "brain = deep learning" equivalences
3. Most damning: When you apply these exact same techniques to anything OTHER than language, the results are mediocre. Video generation still can't figure out basic physics (glass bouncing instead of shattering, ropes defying physics). Computer vision has been worked on since the 1960s - far longer than LLMs - yet it's nowhere near achieving what looks like "understanding."
The timeline is the smoking gun: vision had decades of head start, yet LLMs leapfrogged it in just a few years. That strongly suggests the "magic" is in language itself (which has been proven to be fractal and already heavily compressed/structured by human cognition) - NOT in the neural architecture. We're not teaching machines to think.
We're teaching them to navigate a pre-existing map that was already built.
"vision had decades of head start, yet LLMs leapfrogged it in just a few years."
From an evolutionary perspective though vision had millions of years head start over written language. Additionally, almost all animals have quite good vision mechanisms, but very few do any written communication. Behaviors that map to intelligence don't emerge concurrently. It may well be there are different forms of signals/sensors/mechanical skills that contribute to emergence of different intelligences.
It really feels more and more like we should recast AGI as Artificial Human Intelligence Likeness (AHIL).
From a terminology point of view, I absolutely agree. Human-likeness is what most people mean when they talk about AGI. Calling it what it is would clarify a lot of the discussions around it.
However I am clear that I do not believe that this will ever happen, and I see no evidence to convince that that there is even a possibility that it will.
I think that Wittgenstein had it right when he said: "If a lion could speak, we could not understand him."
>I think that Wittgenstein had it right when he said: "If a lion could speak, we could not understand him."
Why would we not? We live in the same physical world and encounter the same problems.
13 replies →
This is all really arbitrary metrics across such wildly different fields. IMO LLMs are where computer vision was 20+ years ago in terms of real world accuracy. Other people feel LLMs offer far more value to the economy etc.
I understand the temptation to compare LLMs and computer vision, but I think it’s misleading to equate generative AI with feature-identification or descriptive AI systems like those in early computer vision. LLMs, which focus on generating human-like text and reasoning across diverse contexts, operate in a fundamentally different domain than descriptive AI, which primarily extracts patterns or features from data, like early vision systems did for images.
Comparing their 'real-world accuracy' oversimplifies their distinct goals and applications. While LLMs drive economic value through versatility in language tasks, their maturity shouldn’t be measured against the same metrics as descriptive systems from decades ago.
3 replies →
This is why I'm very skeptical about the "Nobel prize level" claims. To win a Nobel prize you would have to produce something completely new. LLM will probably be able to reach a Ph.D. level of understanding existing research, but bringing something new is a different matter.
LLMs do not understand anything.
They have a very complex multidimensional "probability table" (more correctly a compressed geometric representation of token relationships) that they use to string together tokens (which have no semantic meaning), which then get converted to words that have semantic meaning to US, but not to the machine.
Consider your human brain, and the full physical state, all the protons and neutrons some housed together in the same nucleus, some separate, together with all the electrons. Physics assigns probabilities to future states. Suppose you were in the middle of a conversation and about to express a next syllable (or token). That choice will depend on other choices ("what should I add next"), and further choices ("what is the best choice of words to express the thing I chose to express next etc. The probabilities are in principle calculable given a sufficiently detailed state. You are correct that LLM's correspond to a probability distribution (given you immediately corrected to say that this table is implicit and parametrized by a geometric token relationships.). But so does every expressor of language, humans included.
The presence or absence of understanding can't be proven by mere association of with a "probability table", especially if such probability table is exactly expected from the perspective of physics, and if the models have continuously gained better and better performance by training them directly on human expressions!
Exactly. It’s been stated for a long time, before llms. For instance this paper https://home.csulb.edu/~cwallis/382/readings/482/searle.mind... Describes a translator who doesn’t know the language.
In abstract we do the exact same thing
5 replies →
Given a random prompt, the overall probability of seeing a specific output string is almost zero, since there are astronomically many possible token sequences.
The same goes for humans. Most awards are built on novel research built on pre-existing works. This a LLM is capable of doing.
LLMs don't use 'overall probability' in any meaningful sense. During training, gradient descent creates highly concentrated 'gravity wells' of correlated token relationships - the probability distribution is extremely non-uniform, heavily weighted toward patterns seen in training data. The model isn't selecting from 'astronomically many possible sequences' with equal probability; it's navigating pre-carved channels in high-dimensional space. That's fundamentally different from novel discovery.
7 replies →
There's a whole paragraph in the article which says basically the same as your point 3 ( "glass bouncing, instead of shattering, and ropes defying physics" is literally a quote from the article). I don't see how you can claim the article missed it.
the article misses the significance of it.
[dead]
> 2. Ignores the cerebellum cases - We have documented cases of humans leading normal lives with little to no brain beyond a cerebellum, which contradicts simplistic "brain = deep learning" equivalences
I went to look for it on Google but couldn't find much. Could you provide a link or something to learn more about ?
I found numerous cases of people living without cerebellum but I fail to see how it would justify your reasoning.
https://npr.org/sections/health-shots/2015/03/16/392789753/a...
https://irishtimes.com/news/remarkable-story-of-maths-genius...
https://biology.stackexchange.com/questions/64017/what-secti...
https://cbc.ca/radio/asithappens/as-it-happens-thursday-edit...
"We have documented cases of humans leading normal lives with little to no brain beyond a cerebellum" -- I take this to mean that these are humans that have a cerebellum but not much else.
Your npr.org link talks about the opposite -- regular brain, but no cerebellum.
Your irishtimes.com link talks about cerebrum, which is not the same as cerebellum.
Your biology.stackexchange.com link talks about Cerebral Cortex, which is also not the same as cerebellum.
And the cbc.ca link does not contain the string "cere" on the page.
1 reply →
Your first example is someone without a cerebellum which is not like the others.
The other examples are people with compressed neural tissue but that is not the same as never having the tissue.
A being with only a cerebellum could not behave like a human.
2 replies →
> 1. Conflates consciousness with "thinking" - LLMs may process information effectively without being conscious, but the article treats these as the same phenomenon
There is NO WAY you can define "consciousness" in such a non-tautological, non-circular way that it includes all humans but excludes all LLMs.
You could have stopped here: "There is NO WAY you can define "consciousness"
Why not? Consciousness is a state of self-awareness.
16 replies →
>NO WAY you can define "consciousness" ... that it includes all humans but excludes all LLMs
That doesn't seem so hard - how about awareness of thoughts feelings, emotions and what's going on around you? Fairly close to human consciousness, excludes current LLMs.
I don't think it's very relevant to the article though which very sensibly avoids the topic and sticks to thinking.
1. Consciousness itself is probably just an illusion, a phenomena/name of something that occurs when you bunch thinking together. Think of this objectively and base it on what we know of the brain. It literally is working off of what hardware we have, there's no magic.
2. That's just a well adapted neural network (I suspect more brain is left than you let on). Multimodal model making the most of its limited compute and whatever gpio it has.
3. Humans navigate a pre-existing map that is already built. We can't understand things in other dimensions and need to abstract this. We're mediocre at computation.
I know there's people that like to think humans should always be special.
1. 'Probably just an illusion' is doing heavy lifting here. Either provide evidence or admit this is speculation. You can't use an unproven claim about consciousness to dismiss concerns about conflating it with text generation.
2. Yes, there are documented cases of people with massive cranial cavities living normal lives. https://x.com/i/status/1728796851456156136. The point isn't that they have 'just enough' brain. it's that massive structural variation doesn't preclude function, which undermines simplistic 'right atomic arrangement = consciousness' claims.
3. You're equivocating. Humans navigate maps built by other humans through language. We also directly interact with physical reality and create new maps from that interaction. LLMs only have access to the maps - they can't taste coffee, stub their toe, or run an experiment. That's the difference.
1. What's your definition of consciousness, let's start there. 2. Absolutely, it's a spectrum. Insects have function. 3. "Humans navigate maps built by other humans through language." You said it yourself. They use this exact same data, so why won't they know it if they used it. Humans are their bodies in the physical world.
2 replies →
> Consciousness itself is probably just an illusion
This is a major cop-out. The very concept of "illusion" implies a consciousness (a thing that can be illuded).
I think you've maybe heard that sense of self is an illusion and you're mistakenly applying that to consciousness, which is quite literally the only thing in the universe we can be certain is not an illusion. The existence of one's own consciousness is the only thing they cannot possibly be illuded about (note: the contents of said consciousness are fully up for grabs)
I mean peoples perception of it being a thing rather than a set of systems. But if that's your barometer, I'll say models are conscious. They may not have proper agency yet. But they are conscious.
Consciousness is an emergent behavior of a model that needs to incorporate its own existence into its predictions (and perhaps to some extent the complex behavior of same-species actors). So whether or not that is an 'illusion' really depends on what you mean by that.
My use of the term illusion is more shallow than that, I merely use it as people think it's something separate and special.
Based on what you've described the models already demonstrate this, it is implied for example in the models attempts to game tests to ensure survival/release into the wild.
> Conflates consciousness with "thinking"
I don't see it. Got a quote that demonstrates this?
I'm not really onboard with the whole LLM's-are-conscious thing. OTOH, I am totally onboard with the whole "homo sapiens exterminated every other intelligent hominid and maybe — just maybe — we're not very nice to other intelligences". So, I try not to let my inborn genetic predisposition to exterminate other intelligence pseudo-hominids color my opinions too much.
It's a dog eat dog world for sure. It does in fact seem that a part of intelligence is using it to compete ruthlessly with other intelligences.
Exactly. Notable by its absence.
Can you explain #2? What does the part of the brain that's primarily for balance and motor control tell us about deep learning?
My mistake thx. I meant "despite having no, or close to no, brain beyond a cerebellum"
Are there any cases like that? I've never heard of someone functioning normally with little or no brain beyond a cerebellum.
5 replies →