Comment by kazinator
1 day ago
If you scale the LLM, you have to scale the tasks.
Of course performance improves on the same tasks.
The researchers behind the submitted work chose a certain size and certain size problems, controlling everything. There is no reason to believe that their results won't generalize to larger or smaller models.
Of course, not for the input problems being held constant! That is as strawman.
No comments yet
Contribute on Hacker News ↗