Comment by hhjinks
2 years ago
> To broadly be able to predict human speech you need to broadly be able to predict the human mind
This is a non sequitur. The human mind does a whole lot more than string words together. Being able to predict which word would logically follow another does not require the ability to predict anything other than just that.
I think what the commenter is saying is that, in time, language models too will do a lot more than string words together. If it's large enough, and you train it well enough to respond to “what's the best next move in this chess position?” prompts with good moves, it will inevitably learn chess.
I don't think that follows, necessarily. Chess has an unfathomable amount of states. While the LLM might be able to play chess competently, I would not say it has learned chess unless it is able to judge the relative strength of various moves. From my understanding, an LLM will not judge future states of a chess game when responding to such a prompt. Without that ability, it's no different than someone receiving anal bead communications from Magnus Carlsen.
An LLM could theoretically create a model with which to understand chess and predict a next move, you just need to adjust the training data and train the model until that behavior appears.
The expressiveness of language lets this be true of almost everything.
Exactly. Since language is a compressed and transmittable result of our thought, to predict text as accurately as possible requires you do the same. A model with understanding of the human mind will outperform one without.
> Being able to predict which word would logically follow another does not require the ability to predict anything other than just that.
Why? Wouldn't you expect that technique to generally fail if it isn't intelligent enough to know what's happening in the sentence?