Comment by bigstrat2003
10 days ago
> Humans make mistakes all the time and we happily pay for those by the hour as long as the mistakes stay within an acceptable threshold.
We don't, however, continue to pay for the same person who keeps making the same mistakes and doesn't learn from them. Which is what happens with LLMs.
This is why easy "out of the box" continual learning is absolutely essential in practice. It's not like the LLM is incapable of solving tasks, it simply wasn't trained for your specific one. There are optimizers like DSPy that let you validate against a test dataset to increase reliability at the expense of generality.