Just as in Pascal's wager, the conclusion relies on the unwarranted assumption which privileges a particular outcome over its exact opposite - e.g. a diety with exactly inverted criteria for heaven and hell, punishing those who believe in Christian God, and "Roko's antibasilisk", which spares those people who'd get punished by Roko's basilisk and punishes everyone else.
exactly! I gave a heartfelt letter to my shredder the other day and it simply destroyed it. issues like these are why AI alignment research is so critical.
Roko's basilisk is just Pascal's wager for the geeks.
Just as in Pascal's wager, the conclusion relies on the unwarranted assumption which privileges a particular outcome over its exact opposite - e.g. a diety with exactly inverted criteria for heaven and hell, punishing those who believe in Christian God, and "Roko's antibasilisk", which spares those people who'd get punished by Roko's basilisk and punishes everyone else.
Hell, what about Skynet?
and what about my shredder?
exactly! I gave a heartfelt letter to my shredder the other day and it simply destroyed it. issues like these are why AI alignment research is so critical.