{"id":222330,"date":"2025-09-23T12:07:34","date_gmt":"2025-09-23T17:07:34","guid":{"rendered":"https:\/\/lifeboat.com\/blog\/2025\/09\/sorry-mr-yudkowsky-well-build-it-and-everything-will-be-fine"},"modified":"2025-09-23T12:07:34","modified_gmt":"2025-09-23T17:07:34","slug":"sorry-mr-yudkowsky-well-build-it-and-everything-will-be-fine","status":"publish","type":"post","link":"https:\/\/lifeboat.com\/blog\/2025\/09\/sorry-mr-yudkowsky-well-build-it-and-everything-will-be-fine","title":{"rendered":"Sorry Mr. Yudkowsky, we\u2019ll build it and everything will be fine"},"content":{"rendered":"<p><a class=\"aligncenter blog-photo\" href=\"https:\/\/lifeboat.com\/blog.images\/sorry-mr-yudkowsky-well-build-it-and-everything-will-be-fine.jpg\"><\/a><\/p>\n<p>Review of \u201cIf Anyone Builds It, Everyone Dies: Why Superhuman AI Would Kill Us All\u201d (2025), by Eliezer Yudkowsky and Nate Soares, with very critical commentary.<\/p>\n<p><center>I\u2019be been reading the book \u201c<a href=\"https:\/\/www.hachettebookgroup.com\/titles\/eliezer-yudkowsky\/if-anyone-builds-it-everyone-dies\/9780316595667\/\" target=\"_blank\" rel=\"noreferrer noopener\"><em>If Anyone Builds It, Everyone Dies: Why Superhuman AI Would Kill Us All<\/em><\/a>\u201d (2025), by Eliezer Yudkowsky and Nate Soares, published last week.<\/center><\/p>\n<p>Yudkowsky and Soares present a stark warning about the dangers of developing artificial superintelligence (ASI), defined as artificial intelligence (AI) that vastly exceeds human intelligence. The authors argue that creating such AI using current techniques would almost certainly lead to human extinction and emphasize that ASI poses an existential threat to humanity. They argue that the race to build smarter-than-human AI is not an arms race but a \u201csuicide race,\u201d driven by competition and optimism that ignores fundamental risks.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Review of \u201cIf Anyone Builds It, Everyone Dies: Why Superhuman AI Would Kill Us All\u201d (2025), by Eliezer Yudkowsky and Nate Soares, with very critical commentary. I\u2019be been reading the book \u201cIf Anyone Builds It, Everyone Dies: Why Superhuman AI Would Kill Us All\u201d (2025), by Eliezer Yudkowsky and Nate Soares, published last week. Yudkowsky [\u2026]<\/p>\n","protected":false},"author":715,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[12,9,6],"tags":[],"class_list":["post-222330","post","type-post","status-publish","format-standard","hentry","category-existential-risks","category-military","category-robotics-ai"],"_links":{"self":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/222330","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/users\/715"}],"replies":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/comments?post=222330"}],"version-history":[{"count":0,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/222330\/revisions"}],"wp:attachment":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/media?parent=222330"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/categories?post=222330"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/tags?post=222330"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}