← Back to context

Comment by quesera

2 years ago

I was reusing your example in the abstract form.

If the AI does something in the physical world which we do not like, we sever its connection. Unless some people with more power like it more than the rest of us do.

Regarding orthogonal goals: I don't think an AI has goals. Or motivations. Now obviously a lot of destruction can be a side effect, and that's an inherent risk. But it is, I think, a risk of human creation. The AI does not have a survival instinct.

Energy and resources are limiting factors. The first might be solvable! But currently it serves as a failsafe against prolonged activity with which we do not agree.

So I think we have some differences in definition. I am assuming we have an ASI, and then going on from there.

Minimally an ASI (Artificial Super Intelligence) would:

1. Be able to solve all cognitively demanding tasks humans can solve and tasks humans cannot solve (i.e. develop new science), hence "super" intelligent.

2. Be an actively evolving agent (not a large, static compositional function like today's frontier models)

For me intelligence is a problem solving quality of a living thing, hence point 2. I think it might be the case to become super-intelligent, you need to be an agent interfacing with the world, but feel free to disagree here.

Though, if you accept the above formulation of ASI, then by definition (point 2) it would have goals.

Then based on point 1, I think it might not be as simple as "If the AI does something in the physical world which we do not like, we sever its connection."

I think a super-intelligence would be able to perform actions that prevent us from doing that, given that it is clever enough.

  • I agree that the definitions are slippery and evolving.

    But I cannot make the leap from "super intelligent" to "has access to all the levers of social and physical systems control" without the explicit, costly, and ongoing, effort of humans.

    I also struggle with the conflation of "intelligent" and "has free will". Intelligent humans will argue that not even humans have free will. But assuming we do, when our free will contradicts the social structure, society reacts.

    I see no reason to believe that the emergent properties of a highly complex system will include free will. Or curiosity, or a sense of humor. Or a soul. Or goals, or a concept of pleasure or pain, etc. And I think it's possible to be "intelligent" and even "sentient" (whatever that means) without those traits.

    Honestly -- and I'm not making an accusation here(!) -- this fear of AI reminds me of the fear of replacement / status loss. We humans are at the top of the food chain on all scales we can measure, and we don't want to be replaced, or subjugated in the way that we presently subjugate other species.

    This is a reasonable fear! Humans are often difficult to share a planet with. But I don't think it survives rational investigation.

    If I'm wrong, I'll be very very wrong. I don't think it matters though, there is no getting off this train, and maybe there never was. There's a solid argument for being in the engine vs the caboose.

    • Totally fair points.

      > I cannot make the leap from "super intelligent" to "has access to all the levers of social and physical systems control" without the explicit, costly, and ongoing, effort of humans.

      Yeah this is a fair point! The super intellect may just convince humans, which seems feasible. Either way, the claim that there are 0 paths here for a super intelligence is pretty strong so I feel like we can agree on: It'd be tricky, but possible given sufficient cleverness.

      > I see no reason to believe that the emergent properties of a highly complex system will include free will.

      I really do think in the next couple years we will be explicitly implementing agentic architectures in our end-to-end training of frontier models. If that is the case, obviously the result would have something analogous to goals.

      I don't really care about it's phenomenal quality or anything, it's not relevant to my original point.

      2 replies →