← Back to context

Comment by mlsu

6 months ago

Humans have more access to the real world. These models have to tokenize everything and put it into words, but so much information is outside of words. These models may well be super intelligent but their intelligence is locked inside of a cage (the tokenizer).

Even in the world where AI has full control of lights out factories (again, doubt it. something goes wrong at the factory, you gotta send a guy in), human beings still need to look each other in the eye and communicate, they need to touch each other. Not only that, they need to be seen and acknowledged by other human beings.

"AI" cannot ever replace this. People whose intelligence is their pride/identity kind of miss this. Stupid people are capable of loving each other more deeply and more completely than any machine ever will love them.

You basically just said people will be the janitors, the on-site fixers, and the personification of decisions and that they will still be able to live fulfilling lives in the real world. I think that is perfectly in line with what the parent wrote.

All those things could be done by humanoid robots. AI models aren’t limited to words, as we’ve seen with video models. Gpt 4o, which has been out for over a year, is natively multimodal. Robotics companies are training robots to take in all the data they have avaliable, video, audio, and interpret them all together in context. There is the core substrate of tokens, yes, but largely it is just a standard “bit” level of information for AI brains, not some essential limiter that will keep AI from understanding all the soft, abstract stuff that humans can. If you look at o3 now, just feeding it images, it clearly now can reason in a way closer to humans than a calculator is to it.