← Back to context

Comment by cubefox

16 hours ago

> I don't think we have much to worry about in terms of economic disruption. At this point it seems pretty clear that LLMs are having a major impact on how software is built, but for almost every other industry the practical effects are mostly incremental.

You clearly didn't read the post. He is talking about AI that is smarter than any human, not today's LLMs. The fact that powerful AI doesn't exist yet doesn't mean there is nothing to worry about.

> You clearly didn't read the post

This kind of petty remark is like a reverse em dash. Greetings fellow human.

Anyway, I did read it. The author's description of a future AI is basically just a more advanced version of LLMs

> By “powerful AI,” I have in mind an AI model—likely similar to today’s LLMs in form, though it might be based on a different architecture, might involve several interacting models, and might be trained differently—with the following properties:

They then go on to list several properties that meet their definition, but what I'm trying to explain in my comment is that I don't accept them all at face value. I think it's fair to critique from that perspective since the author explicitly modeled their future based on today's LLMs, unlike many AI essays that skip straight to the super intelligence meme as their premise.

  • > They then go on to list several properties that meet their definition

    No, these properties are part of his definition. To say that we have nothing to worry about because today's LLMs don't have these properties misses the point.