Comment by Skyy93

5 months ago

Let's just assume the model is a statistical parrot, which it probably is. The probability for the next token is influenced based on the input. So far so good, if I now ask a question, the probability that I generate the corresponding answer increases. But is it the right one? This is exactly where CoT tries to start, in which context is generated you change the probability of the tokens for the answer and we can at least experimentally show that the answers get better. Perhaps it is easier to speak of a kind of refinement, the more context is generated, the more focused the model is on the currently important topic.