{"id":202572,"date":"2024-12-27T12:43:28","date_gmt":"2024-12-27T18:43:28","guid":{"rendered":"https:\/\/lifeboat.com\/blog\/2024\/12\/ais-achilles-heel-researchers-expose-major-model-security-flaw"},"modified":"2024-12-27T12:43:28","modified_gmt":"2024-12-27T18:43:28","slug":"ais-achilles-heel-researchers-expose-major-model-security-flaw","status":"publish","type":"post","link":"https:\/\/lifeboat.com\/blog\/2024\/12\/ais-achilles-heel-researchers-expose-major-model-security-flaw","title":{"rendered":"AI\u2019s Achilles\u2019 Heel: Researchers Expose Major Model Security Flaw"},"content":{"rendered":"<p><a class=\"aligncenter blog-photo\" href=\"https:\/\/lifeboat.com\/blog.images\/ais-achilles-heel-researchers-expose-major-model-security-flaw2.jpg\"><\/a><\/p>\n<p><strong>Researchers used electromagnetic signals to steal and replicate AI models from a Google Edge TPU with 99.91% accuracy, exposing significant vulnerabilities in AI systems and calling for urgent protective measures.<\/strong><\/p>\n<p>Researchers have shown that it\u2019s possible to steal an artificial intelligence (AI) model without directly hacking the device it runs on. This innovative technique requires no prior knowledge of the software or architecture supporting the AI, making it a significant advancement in model extraction methods.<\/p>\n<p>\u201cAI models are valuable, we don\u2019t want people to steal them,\u201d says Aydin Aysu, co-author of a paper on the work and an associate professor of electrical and computer engineering at <a href=\"https:\/\/scitechdaily.com\/tag\/north-carolina-state-university\/\"> North Carolina State University<\/a>. \u201cBuilding a model is expensive and requires significant computing sources. But just as importantly, when a model is leaked, or stolen, the model also becomes more vulnerable to attacks \u2013 because third parties can study the model and identify any weaknesses.\u201d<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Researchers used electromagnetic signals to steal and replicate AI models from a Google Edge TPU with 99.91% accuracy, exposing significant vulnerabilities in AI systems and calling for urgent protective measures. Researchers have shown that it\u2019s possible to steal an artificial intelligence (AI) model without directly hacking the device it runs on. This innovative technique requires [\u2026]<\/p>\n","protected":false},"author":396,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[6,1492],"tags":[],"class_list":["post-202572","post","type-post","status-publish","format-standard","hentry","category-robotics-ai","category-security"],"_links":{"self":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/202572","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/users\/396"}],"replies":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/comments?post=202572"}],"version-history":[{"count":0,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/202572\/revisions"}],"wp:attachment":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/media?parent=202572"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/categories?post=202572"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/tags?post=202572"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}