Comment by specialist
2 days ago
u/ipaddr is probably referring to
1) the dearth of new (novel) training data. Hence the mad scramble to hoover up, buy, steal, any potentially plausible new sources.
2) diminishing returns of embiggening compute clusters for training LLMs and size of their foundation models.
(As you know) You're referring to Wright's Law aka experience learning curve.
So there's a tension.
Some concerns that we're nearing the ceiling for training.
While the cost of applications using foundation models (implementing inference engines) is decreasing.
Someone smarter than me will have to provide the slopes of the (misc) learning curves.
I was not aware of (or had forgotten) the term "Wright's law" [1], but that indeed is what I was thinking of. It looks like some may use the term "learning curve" to refer to the same idea (efficiency gains that follow investment); the Wikipedia page on "Learning curve" [2] includes references to Wright.
[1] https://en.wikipedia.org/wiki/Experience_curve_effect
[2] https://en.wikipedia.org/wiki/Learning_curve