Comment by JKCalhoun
3 hours ago
Interesting to me, during that crazy period when Sutskever ultimately ended up leaving OpenAI, I thought perhaps he had shot himself in the foot to some degree (not that I have any insider information—just playing stupid observer from the outside).
The feeling I have now is that it was a fine decision for him to have made. It made a point at the time, perhaps moral, perhaps political. And now it seems, despite whatever cost there was for him at the time, the "golden years" of OpenAI (and LLM's in general) may have been over anyway.
To be sure, I happen to believe there is a lot of mileage for LLMs even in their current state—a lot of use-cases, integration we have yet to explore. But Sutskever I assume is a researcher and not a plumber—for him the LLM was probably over.
One wonders how long before one of these "break throughs". On one hand, they may come about serendipitously, and serendipity has no schedule. It harkens back to when A.I. itself was always "a decade away". You know, since the 1950's or so.
On the other hand, there are a lot more eyeballs on AI these days than there ever were in Minsky's* day.
(*Hate to even mention the man's name these days.)
> To be sure, I happen to believe there is a lot of mileage for LLMs even in their current state—a lot of use-cases, integration we have yet to explore. But Sutskever I assume is a researcher and not a plumber—for him the LLM was probably over.
Indeed. Humans are a sucker for a quick answer delivered confidently. And The industry coalesced around LLM's once it was able to output competent, confident, corporate (aka HR-approved) english, which for many AI/DL/ML/NN researchers was actually a bit of a bummer. Reason I say that is because that milestone suddenly made the "[AGI is] always a decade away" to seeming much more imminent. Thus the focus of investment in the space shifted from actual ML/DL/NN research to who could convert largest pile of speculatively leveraged money into pallets of GPU's and data to feed them as "throw more compute/data" at it was a quicker/more reliable way to realize performance gains than investing in research did. Yes, research would inevitably yield results, but it's incredibly hard to forceast how long it takes for research to yield tangible results and harder still to quantify that X dollars will result in Y result in Z time compared to X dollars buys Y compute deployed in Z time. With the immense speculative backed FOMO and the potential valuation/investment that could result from being "the leader" in any given regard, it's no wonder that BigTech chose to primarily invest in the latter, thus leaving to those working in the former space to start considering looking elsewhere to continue actual research.