← Back to context

Comment by killerstorm

4 hours ago

What does LLM need to do for you to consider it "smart"?

To me they seem to be pretty damn smart, to put it mildly. They sometimes do stupid things - but so do smart people!

Not OP, but I think the argument here would be not that LLMs "are not smart" but that smart is just the wrong category of thing to describe an LLM as.

A calculator can do very complex sums very quickly, but we don't tend to call it "smart" because we don't think it's operating intelligently to some internal model of the world. I think the "LLMs are AGI" crowd would say that LLMs are, but it's perfectly consistent to think the output of LLMs is consistent/impressive/useful, but still maintain that they aren't "smart" in any meaningful way.

  • > "we don't think it's operating intelligently to some internal model of the world"

    Okay, but you have to actually address why you think LLMs lack an "internal model of the world"

    You can train one on 1930s text, and then teach it Python in-context.

    They've produced multiple novel mathematical proofs now; Terrance Tao is impressed with them as research assistants.

    You can very clearly ask them questions about the world, and they'll produce answers that match what you'd get from a "model" of the world.

    What are weights, if not a model of the world? It's got a very skewed perspective, certainly, since it's terminally online and has never touched grass, but it still very clearly has a model of the world.

    I'd dare say it's probably a more accurate model than the average person has, too, thanks to having Wikipedia and such baked in.

> To me they seem to be pretty damn smart

That's the sorcery mentioned in the GP, the issue comes when people believe it to be smart however in reality it is just a next word prediction. Gives the impression it's actually thinking, and this is by design. Personally I think it's dangerous in the sense it gives users a false sense of confidence in the LLM and so a LOT of people will blindly trust it. This isn't a good thing.

  • I'm curious how you think "word predictor" meaningfully describes an instruct model that has developed novel mathematical proofs that have eluded mathematicians for decades?

    edit:

    You cannot predict all the actions or words of someone smarter than you. If I could always predict Magnus Carlsen's next chess move, I'd be at least as good at chess as Magnus - and that would have to involve a deep understanding of chess, even if I can't explain my understanding.

    I can't predict the next token in a novel mathematical proof unless I've already understood the solution.

    • I think that's more of a limitation in how people think about word predictors

      If you can predict the words a bright person will say about X... Isn't that some truly astounding tool? That could be used in myriad useful ways if one is a little creative with it

      Since it's also "alien" it can also detect and explore paths that we simply haven't noticed since their biases aren't quite the same as ours

      1 reply →

  • What's the difference between "smart" and "next word prediction", at this point? Back when they first came out, sure, but now they can write code and create art.

    What would it take for you to concede a future model was smart?

    • My personal take would always be that it produces something that isn't in the training set, ie: Demonstrable Creativity, or innovation.

      For example, it's training set it purely engineering and code with general language data set, would be "aware" what art is, but has never seen an artistic image, aware what colours are and able to create something it never saw before.

      Like a child with a paintbrush, there is an intuitive behavior that happens.

      1 reply →

They aren’t smart, they approximate language constructs. They don’t have believes, ideas, etc. but have a few rounds of discussions with any LLMs and you see how they are probabilistic autocompletes based on whatever patterns from rounds of discussions you feed them

  • At what point does autocomplete stop being "just autocomplete"?

    Clearly there's a limit. For example, if an alien autocomplete implementation were to fall out of a wormhole that somehow manages to, say, accurately complete sentences like "S&P 500, <tomorrow's date>:" with tomorrow's actual closing value today, I'd call that something else.

    • You can call it however you want. The point of using the term autocomplete is to make the underlying technology relatable and remove the mystic from it. In any case, your alien autocomplete wouldn’t be an LLM if it can predict the future

      > At what point does autocomplete stop being "just autocomplete"?

      Every single discussion on the internet is a repeat of https://en.wikipedia.org/wiki/Loki%27s_wager it seems…

Are they smart or are they imitating things smart people did? (and if so, is there a difference?)

LLMs are amazing. You can call them 'smart', but they're not intelligent and never will be.

They are useful but a cul de sac for heading toward AGI.

  • You can always redefine "intelligent" so that humans meet the requirements but AIs don't.

    A better model to use is this: LLMs possess a different type of intelligence than us, just like an intelligent alien species from another planet might.

    A calculator has a very narrow sort of intelligence. It has near perfect capability in a subset of algebra with finite precision numbers, but that's it.

    An old-school expert system has its own kind of intelligence, albeit brittle and limited to the scope of its pre-programmed if-then-else statements.

    By extension, an AI chat bot has a type of intelligence too. Not the same as ours, but in many ways superior, just as how a calculator is superior to a human at basic numeric algebra. We make mistakes, the calculator does not. We make grammar and syntax errors all the time, the AI chat bots generally never do. We speak at most half a dozen languages fluently, the chat bots over a hundred. We're experts in at most a couple of fields of study, the chat bots have a very wide but shallow understanding. Etc.

    Don't be so narrow minded! Start viewing all machines (and creatures) as having some type of intelligence instead of a boolean "have" or "have not" intelligence.

    • > A calculator has a very narrow sort of intelligence.

      Have you ever heard anyone refer to a calculator as intelligent?

      These companies have a vested interest in making the product appear more human/smart than it is. It's new tech smeared with the same ole marketing matter.

    • Would you say that a display and a printer are a perfect painter because they can render images? And a speaker is a very good musician because they can produce sound?

      The LLM tasks is to produce a string of words according to an internal model trained on texts written by humans (and now generted by other LLMs). This is not intelligence.

      2 replies →