Comment by feisty0630
3 days ago
Interesting that it reads a bit like it came from a Markov chain rather than an LLM. Perhaps limited training data?
3 days ago
Interesting that it reads a bit like it came from a Markov chain rather than an LLM. Perhaps limited training data?
Early LLMs used to have this often. I think's that where the "repetition penalty" parameter comes from. I suspect output quality can be improved with better sampling parameters.
It is lacking all recorded text from the past 200 years. ;)
It would be interesting to know how much text was generated per century!