{"id":176025,"date":"2023-11-14T13:24:23","date_gmt":"2023-11-14T19:24:23","guid":{"rendered":"https:\/\/lifeboat.com\/blog\/2023\/11\/glasses-use-sonar-ai-to-interpret-upper-body-poses-in-3d"},"modified":"2023-11-14T13:24:23","modified_gmt":"2023-11-14T19:24:23","slug":"glasses-use-sonar-ai-to-interpret-upper-body-poses-in-3d","status":"publish","type":"post","link":"https:\/\/lifeboat.com\/blog\/2023\/11\/glasses-use-sonar-ai-to-interpret-upper-body-poses-in-3d","title":{"rendered":"Glasses use sonar, AI to interpret upper body poses in 3D"},"content":{"rendered":"<p><a class=\"aligncenter blog-photo\" href=\"https:\/\/lifeboat.com\/blog.images\/glasses-use-sonar-ai-to-interpret-upper-body-poses-in-3d.jpg\"><\/a><\/p>\n<p>Throughout history, sonar\u2019s distinctive \u201cping\u201d has been used to map oceans, spot enemy submarines and find sunken ships. Today, a variation of that technology \u2013 in miniature form, developed by Cornell researchers \u2013 is proving a game-changer in wearable body-sensing technology.<\/p>\n<p>PoseSonic is the latest sonar-equipped wearable from Cornell\u2019s <a href=\"https:\/\/www.scifilab.org\">Smart Computer Interfaces for Future Interactions (SciFi) lab<\/a>. It consists of off-the-shelf eyeglasses outfitted with micro sonar that can track the wearer\u2019s upper body movements in 3D through a combination of inaudible soundwaves and artificial intelligence (AI).<\/p>\n<p>With further development, PoseSonic could enhance augmented reality and virtual reality, and track detailed physical and behavioral data for personal health, the researchers said.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Throughout history, sonar\u2019s distinctive \u201cping\u201d has been used to map oceans, spot enemy submarines and find sunken ships. Today, a variation of that technology \u2013 in miniature form, developed by Cornell researchers \u2013 is proving a game-changer in wearable body-sensing technology. PoseSonic is the latest sonar-equipped wearable from Cornell\u2019s Smart Computer Interfaces for Future Interactions [\u2026]<\/p>\n","protected":false},"author":662,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1498,1495,6,1879,1977],"tags":[],"class_list":["post-176025","post","type-post","status-publish","format-standard","hentry","category-augmented-reality","category-health","category-robotics-ai","category-virtual-reality","category-wearables"],"_links":{"self":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/176025","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/users\/662"}],"replies":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/comments?post=176025"}],"version-history":[{"count":0,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/176025\/revisions"}],"wp:attachment":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/media?parent=176025"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/categories?post=176025"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/tags?post=176025"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}