{"id":222686,"date":"2025-09-30T04:34:50","date_gmt":"2025-09-30T09:34:50","guid":{"rendered":"https:\/\/lifeboat.com\/blog\/2025\/09\/is-violent-ai-human-conflict-inevitable"},"modified":"2025-09-30T04:34:50","modified_gmt":"2025-09-30T09:34:50","slug":"is-violent-ai-human-conflict-inevitable","status":"publish","type":"post","link":"https:\/\/lifeboat.com\/blog\/2025\/09\/is-violent-ai-human-conflict-inevitable","title":{"rendered":"Is violent AI-human conflict inevitable?"},"content":{"rendered":"<p><a class=\"aligncenter blog-photo\" href=\"https:\/\/lifeboat.com\/blog.images\/is-violent-ai-human-conflict-inevitable.jpg\"><\/a><\/p>\n<p>Are you worried that artificial intelligence and humans will go to war? AI experts are. In 2023, a group of elite thinkers signed onto the <a href=\"https:\/\/safe.ai\/work\/statement-on-ai-risk\" target=\"_blank\">Center for AI Safety\u2019s<\/a> statement that \u201cMitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.\u201d<\/p>\n<p>In a survey <a href=\"https:\/\/aiimpacts.org\/wp-content\/uploads\/2023\/04\/Thousands_of_AI_authors_on_the_future_of_AI.pdf\" target=\"_blank\">published in 2024<\/a>, 38% to 51% of top-tier AI researchers assigned a probability of at least 10% to the statement \u201cadvanced AI leading to outcomes as bad as human extinction.\u201d<\/p>\n<p>The worry is not about the Large Language Models (LLMs) of today, which are essentially huge autocomplete machines, but about <a href=\"https:\/\/builtin.com\/artificial-intelligence\/artificial-general-intelligence\" target=\"_blank\">Advanced General Intelligence<\/a> (AGI)\u2014still hypothetical long-term planning agents that can substitute for human labor across a wide range of society\u2019s economic systems.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Are you worried that artificial intelligence and humans will go to war? AI experts are. In 2023, a group of elite thinkers signed onto the Center for AI Safety\u2019s statement that \u201cMitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.\u201d In a [\u2026]<\/p>\n","protected":false},"author":427,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[11,39,12,6],"tags":[],"class_list":["post-222686","post","type-post","status-publish","format-standard","hentry","category-biotech-medical","category-economics","category-existential-risks","category-robotics-ai"],"_links":{"self":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/222686","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/users\/427"}],"replies":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/comments?post=222686"}],"version-history":[{"count":0,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/222686\/revisions"}],"wp:attachment":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/media?parent=222686"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/categories?post=222686"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/tags?post=222686"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}