{"id":121159,"date":"2021-04-02T01:23:33","date_gmt":"2021-04-02T08:23:33","guid":{"rendered":"https:\/\/lifeboat.com\/blog\/2021\/04\/a-robot-that-senses-hidden-objects"},"modified":"2021-04-02T01:23:33","modified_gmt":"2021-04-02T08:23:33","slug":"a-robot-that-senses-hidden-objects","status":"publish","type":"post","link":"https:\/\/lifeboat.com\/blog\/2021\/04\/a-robot-that-senses-hidden-objects","title":{"rendered":"A robot that senses hidden objects"},"content":{"rendered":"<p><\/p>\n<p><iframe style=\"display: block; margin: 0 auto; width: 100%; aspect-ratio: 4\/3; object-fit: contain;\" src=\"https:\/\/www.youtube.com\/embed\/ZAzeYPcTM78?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; encrypted-media; gyroscope;\n   picture-in-picture\" allowfullscreen><\/iframe><\/p>\n<p>In recent years, robots have gained artificial vision, touch, and even smell. \u201cResearchers have been giving robots human-like perception,\u201d says MIT Associate Professor Fadel Adib. In a new paper, Adib\u2019s team is pushing the technology a step further. \u201cWe\u2019re trying to give robots superhuman perception,\u201d he says.<\/p>\n<p>The researchers have developed a <a href=\"https:\/\/techxplore.com\/tags\/robot\/\" rel=\"tag\" class=\"\">robot<\/a> that uses radio waves, which can pass through walls, to sense occluded objects. The robot, called RF-Grasp, combines this powerful sensing with more traditional computer vision to locate and grasp items that might otherwise be blocked from view. The advance could one day streamline e-commerce fulfillment in warehouses or help a machine pluck a screwdriver from a jumbled toolkit.<\/p>\n<p>The research will be presented in May at the IEEE International Conference on Robotics and Automation. The paper\u2019s lead author is Tara Boroushaki, a research assistant in the Signal Kinetics Group at the MIT Media Lab. Her MIT co-authors include Adib, who is the director of the Signal Kinetics Group; and Alberto Rodriguez, the Class of 1957 Associate Professor in the Department of Mechanical Engineering. Other co-authors include Junshan Leng, a research engineer at Harvard University, and Ian Clester, a Ph.D. student at Georgia Tech.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>In recent years, robots have gained artificial vision, touch, and even smell. \u201cResearchers have been giving robots human-like perception,\u201d says MIT Associate Professor Fadel Adib. In a new paper, Adib\u2019s team is pushing the technology a step further. \u201cWe\u2019re trying to give robots superhuman perception,\u201d he says. The researchers have developed a robot that uses [\u2026]<\/p>\n","protected":false},"author":396,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[6],"tags":[],"class_list":["post-121159","post","type-post","status-publish","format-standard","hentry","category-robotics-ai"],"_links":{"self":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/121159","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/users\/396"}],"replies":[{"embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/comments?post=121159"}],"version-history":[{"count":0,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/posts\/121159\/revisions"}],"wp:attachment":[{"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/media?parent=121159"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/categories?post=121159"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/lifeboat.com\/blog\/wp-json\/wp\/v2\/tags?post=121159"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}