Source: Wired
by Steven Levy
“Philosopher Nick Bostrom recently posted a paper, where he postulated that a small chance of AI annihilating all humans might be worth the risk, because advanced AI might relieve humanity of ‘its universal death sentence.’ That upbeat gamble is quite a leap from his previous dark musings on AI, which made him a doomer godfather. … His more recent book, Deep Utopia, reflects a shift in his focus. Bostrom, who leads Oxford’s Future of Humanity Institute, dwells on the ‘solved world’ that comes if we get AI right.” (05/08/26)