← Back to context

Comment by gmuslera

3 months ago

They probably will fall fast into tragedy of the commons kind of situations. We developed most of our civilization while there was enough room for growing and big decisions were centralized, and started to get into bad troubles when things became global enough.

With AIs some of those "protections" may not be there. And hardcoding strategies to avoid this may already put a limit on what we are simulating.

> We developed most of our civilization while there was enough room for growing and big decisions were centralized, and started to get into bad troubles when things became global enough.

Citation needed. But even if I will get on board with you on that, wouldn't it be to start developing for global scale right from the start, instead of starting in small local islands and then try to rework that into global ecosystem?

  • The problem with emulations is human patience. If you don't need/have human interaction this may run pretty fast. And at the end, what matter is how sustainable it is in the long run.

Does this mean that individual complexity is a natural enemy of group cohesiveness? Or is individual 'selfishness' more a product of evolutionary background.

On our planet we don't have ant colony dynamics at the physical scale of high intelligence (that I know of), but there are very physical limitations to things like food sources.

Virtual simulations don't have the same limitations, so the priors may be quite different.

  • Taking the "best" course of action from your own point of view could not be so good from a more broad perspective. We might have evolved some small group collaboration approaches that in the long run plays better, but in large groups that doesn't go that well. And for AIs trying to optimize something without some big picture vision, things may go wrong faster.

    • They might, but if youre simulating a group you can count group survival as part of the selection criteria.