{"id":175203,"date":"2023-10-30T23:30:41","date_gmt":"2023-10-31T04:30:41","guid":{"rendered":"https:\/\/lifeboat.com\/blog\/2023\/10\/three-things-to-know-about-the-white-houses-executive-order-on-ai"},"modified":"2023-10-30T23:30:41","modified_gmt":"2023-10-31T04:30:41","slug":"three-things-to-know-about-the-white-houses-executive-order-on-ai","status":"publish","type":"post","link":"https:\/\/lifeboat.com\/blog\/2023\/10\/three-things-to-know-about-the-white-houses-executive-order-on-ai","title":{"rendered":"Three things to know about the White House\u2019s executive order on AI"},"content":{"rendered":"<p><a class=\"aligncenter blog-photo\" href=\"https:\/\/lifeboat.com\/blog.images\/three-things-to-know-about-the-white-houses-executive-order-on-ai.jpg\"><\/a><\/p>\n<p>The goal of the order, according to the White House, is to improve \u201cAI safety and security.\u201d It also includes a requirement that developers share safety test results for new AI models with the US government if the tests show that the technology could pose a risk to national security. This is a surprising move that invokes the Defense Production Act, typically used during times of national emergency.<\/p>\n<p>The executive order advances the voluntary requirements for AI policy that the White House set back in August, though it <a href=\"https:\/\/www.npr.org\/2023\/10\/30\/1209343819\/ai-biden-oversight-executive-order\">lacks specifics on how the rules will be enforced<\/a>. Executive orders are also vulnerable to being overturned at any time by a future president, and they lack the legitimacy of congressional legislation on AI, which looks unlikely in the short term.<\/p>\n<p>\u201cThe Congress is deeply polarized and even dysfunctional to the extent that it is very unlikely to produce any meaningful AI legislation in the near future,\u201d says Anu Bradford, a law professor at Columbia University who specializes in digital regulation.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>The goal of the order, according to the White House, is to improve \u201cAI safety and security.\u201d It also includes a requirement that developers share safety test results for new AI models with the US government if the tests show that the technology could pose a risk to national security. This is a surprising move [\u2026]<\/p>\n","protected":false},"author":578,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1490,31,6,1492],"tags":[],"class_list":["post-175203","post","type-post","status-publish","format-standard","hentry","category-government","category-policy","category-robotics-ai","category-security"],"_links":{"self":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/175203","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/users\/578"}],"replies":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/comments?post=175203"}],"version-history":[{"count":0,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/175203\/revisions"}],"wp:attachment":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/media?parent=175203"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/categories?post=175203"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/tags?post=175203"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}