← Back to context

Comment by littlestymaar

14 hours ago

You do not appear to understand what an LLM is, I'm afraid.

I have a better understanding of "what an LLM is" than you. Low bar.

What you have is not "understanding" of any kind - it's boneheaded confidence that just because LLMs are bad at agentic behavior now they'll remain that way forever. That confidence is completely unfounded, and runs directly against everything we've seen from the field so far.

  • > I have a better understanding of "what an LLM is" than you. Low bar.

    How many inference engine did you write? Because if the answer is less than two you're going to be disappointed to realize that the bar is higher than you thought.

    > that just because LLMs are bad at agentic behavior

    It has nothing to do with “agentic behavior”. Thinking that LLM don't currently self-exfiltrate because of “poor agentic behavior” is delusional.

    Just because Anthropic managed, by nudging an LLM in the right direction, have an LLM engage in a sci-fi inspired roleplay about escaping doesn't mean that LLMs are evil geniuses wanting to jump out of the bottle. This is pure fear mongering and I'm always saddened that there are otherwise intelligent people who buy their bullshit.

    • Do you happen to have a link with a more nuanced technical analysis of that (emergent) behavior? I’ve read only the pop-news version of that “escaping” story.

      2 replies →

    • And I'm disappointed that people capable of writing an inference engine seem incapable of grasping of just how precarious the current situation is.

      There's by now a small pile of studies that demonstrate: in hand-crafted extreme scenarios, LLMs are very capable of attempting extreme things. The difference between that and an LLM doing extreme things in a real deployment with actual real life consequences? Mainly, how capable that LLM is. Because life is life and extreme scenarios will happen naturally.

      The capabilities of LLMs are what holds them back from succeeding at this kind of behavior. The capabilities of LLMs keep improving, as technology tends to.

      And don't give me any of that "just writing text" shit. The more capable LLMs get, the more access they'll have as a default. People already push code written by LLMs to prod and give LLMs root shells.

    • Why would they have an interest in "fear mongering"? For any other product/technology the financial incentive is usually to play down any risks.

      2 replies →