← Back to context

Comment by jjk166

6 days ago

Pretty big jump from "we don't yet know the best ways to use this new tool" to "the tool doesn't work well"

I think that's adequately addressed in the article:

> "The other way to look at this is like there's no free lunch here," said Smiley. "We know what the limitations of the model are. It's hard to teach them new facts. It's hard to reliably retrieve facts. The forward pass through the neural nets is non-deterministic, especially when you have reasoning models that engage an internal monologue to increase the efficiency of next token prediction, meaning you're going to get a different answer every time, right? That monologue is going to be different.

> "And they have no inductive reasoning capabilities. A model cannot check its own work. It doesn't know if the answer it gave you is right. Those are foundational problems no one has solved in LLM technology. And you want to tell me that's not going to manifest in code quality problems? Of course it's going to manifest."

You can argue with specifics in there, but they made their case.