Why would a misaligned superintelligence kill us?

While AI is unlikely to be malevolent towards humanity, we might still die as a result of the AI doing instrumental reasoning, with our deaths being either 1) an intentional goal or 2) a side-effect of some other goal:

The default outcome of these is probably not just death on a large scale, but human extinction.

On the other hand, keeping humans around would take only a small fraction of a superintelligence’s resources. Some have argued that an AI might be willing to pay that small cost to keep us around, either if it’s only mostly misaligned and cares about us a little bit, or for various decision-theoretic reasons. That could look like anything from giving us free rein over a small part of the universe to putting us into a kind of zoo. Others think it will be unwilling to pay even that cost. And even if humanity survives like this, that’s not an ideal outcome: many of us might still die, survivors might not like their situation, and most of the universe would be outside of humanity’s reach forever.


  1. But any superintelligence that was both powerful and misaligned enough to consider taking humans apart for their atoms would be modifying the rest of the world radically enough to make human life impossible, anyway. ↩︎



AISafety.info

We’re a global team of specialists and volunteers from various backgrounds who want to ensure that the effects of future AI are beneficial rather than catastrophic.