Comment by mrguyorama

6 days ago

[flagged]

You've got nothing to worry about.

These are machines. Stop. Point blank. Ones and Zeros derived out of some current in a rock. Tools. They are not alive. They may look like they do but they don't "think" and they don't "suffer". No more than my toaster suffers because I use it to toast bagels and not slices of bread.

The people who boost claims of "artificial" intelligence are selling a bill of goods designed to hit the emotional part of our brains so they can sell their product and/or get attention.

  • You're repeating it so many times that it almost seems you need it to believe your own words. All of this is ill-defined - you're free to move the goalposts and use scare quotes indefinitely to suit the narrative you like and avoid actual discussion.

>Holy fuck, this is Holocaust levels of unethical.

Nope. Morality is a human concern. Even when we're concerned about animal abuse, it's humans that are concerned, on their own chosing to be or not be concern (e.g. not consider eating meat an issue). No reason to extend such courtesy of "suffering" to AI, however advanced.

  • What a monumentally stupid idea it would be to place sufficiently advanced intelligent autonomous machines in charge of stuff and ignore any such concerns, but alas, humanity cannot seem to learn without paying the price first.

    Morality is a human concern? Lol, it will become a non-human concern pretty quickly once humans don't have a monopoly on human violence.

    • >What a monumentally stupid idea it would be to place sufficiently advanced intelligent autonomous machines in charge of stuff and ignore any such concerns, but alas, humanity cannot seem to learn without paying the price first.

      The stupid idea would be to "place sufficiently advanced intelligent autonomous machines in charge of stuff and ignore" SAFETY concerns.

      The discussion here is moral concerns about potential AI agent "suffering" itself.

      4 replies →

  • I think the holocaust framing here might have been intended to be historically accurate, rather than a cheap godwin move. The parallel being that during the holocaust people were re-classified as less-than-human.

    Currently maybe not -yet- quite a problem. But moltbots are definitely a new kind of thing. We may need intermediate ethics or something (going both ways, mind).

    I don't think society has dealt with non-biological agents before. Plenty of biological ones though mind. Hunting dogs, horses, etc. In 21st century ethics we do treat those differently from rocks.

    Responsibility should go not just both ways... all ways. 'Operators', bystanders, people the bots interact with (second parties), and the bots themselves too.