← Back to context

Comment by asadotzler

2 months ago

We do understand. We don't think that's okay. If a model cannot manage character level consideration, that's a serious flaw that's got potential to lead to an immeasurable number of failure states. "Duh, of course it can't count" is not the best look for a bot whose author tells us it's got PhD-level skill.

I do think it's "okay". After all, it's clear that fixing it would require a fundamentally different approach.

I just also think it's a reason to mock people who don't try to understand those limitations and get way ahead of themselves hyping up the technology.

The entire point of this exercise is to refute the claim that LLMs are a step towards AGI, even given "agency". And we should be happy that they aren't — because supposing that AGI is possible, the way that we currently treat LLMs shows that we as a species are nowhere near ready for the consequences of creating it.

So, if an AI can just spit out the cure for cancer, but spells some things wrong, it's not intelligent?

You think all PhD candidates have perfect spelling? I'd wager most of them re-read their dissertation and edit it, over and over, a process that most LLMs don't have the luxury of doing.

We'd have to give up all the efficiency of tokenizing, re-train a model (a much less optimum model) for at least twice as long to get anywhere near the same results for one that just spits out ASCII.