{"id":6632,"date":"2023-07-10T18:30:00","date_gmt":"2023-07-10T18:30:00","guid":{"rendered":"https:\/\/www.aiproblog.com\/index.php\/2023\/07\/10\/3-questions-honing-robot-perception-and-mapping\/"},"modified":"2023-07-10T18:30:00","modified_gmt":"2023-07-10T18:30:00","slug":"3-questions-honing-robot-perception-and-mapping","status":"publish","type":"post","link":"https:\/\/www.aiproblog.com\/index.php\/2023\/07\/10\/3-questions-honing-robot-perception-and-mapping\/","title":{"rendered":"3 Questions: Honing robot perception and mapping"},"content":{"rendered":"<p>Author: Madeleine Turner | MIT Laboratory for Information and Decision Systems<\/p>\n<div>\n<p><em>Walking to a friend\u2019s house or browsing the aisles of a grocery store might feel like simple tasks, but they in fact require sophisticated capabilities. That&#8217;s because humans are able to effortlessly understand their surroundings and detect complex information about patterns, objects, and their own location in the environment.<\/em><\/p>\n<p><em>What if robots could perceive their environment in a similar way? That question is on the minds of MIT Laboratory for Information and Decision Systems (LIDS) researchers Luca Carlone and Jonathan How. In 2020, a team led by Carlone released the first iteration of <a href=\"https:\/\/github.com\/MIT-SPARK\/Kimera\">Kimera<\/a>, an open-source library that enables a single robot to construct a three-dimensional map of its environment in real time, while labeling different objects in view. Last year, Carlone\u2019s and How\u2019s research groups (<a href=\"http:\/\/web.mit.edu\/sparklab\/\">SPARK Lab<\/a> and<a href=\"https:\/\/acl.mit.edu\/\"> Aerospace Controls Lab<\/a>) introduced <a href=\"https:\/\/github.com\/MIT-SPARK\/Kimera-Multi\">Kimera-Multi<\/a>, an updated system in which multiple robots communicate among themselves in order to create a unified map. A 2022 <a href=\"https:\/\/ieeexplore.ieee.org\/stamp\/stamp.jsp?arnumber=9686955\" target=\"_blank\" rel=\"noopener\">paper<\/a> associated with the project recently received this year\u2019s <\/em>IEEE Transactions on Robotics<em> King-Sun Fu Memorial Best Paper Award, given to the best paper published in the journal in 2022.<\/em><\/p>\n<p><em>Carlone, who is the Leonardo Career Development Associate Professor of Aeronautics and Astronautics, and How, the Richard Cockburn Maclaurin Professor in Aeronautics and Astronautics, spoke to LIDS about Kimera-Multi and the future of how robots might perceive and interact with their environment.<\/em><\/p>\n<p><strong>Q: <\/strong>Currently your labs are focused on increasing the number of robots that can work together in order to generate 3D maps of the environment. What are some potential advantages to scaling this system?<\/p>\n<p><strong>How: <\/strong>The key benefit hinges on consistency, in the sense that a robot can create an independent map, and that map is self-consistent but not globally consistent. We\u2019re aiming for the team to have a consistent map of the world; that\u2019s the key difference in trying to form a consensus between robots as opposed to mapping independently.<\/p>\n<p><strong>Carlone: <\/strong>In many scenarios it\u2019s also good to have a bit of redundancy. For example, if we deploy a single robot in a search-and-rescue mission, and something happens to that robot, it would fail to find the survivors. If multiple robots are doing the exploring, there\u2019s a much better chance of success. Scaling up the team of robots also means that any given task may be completed in a shorter amount of time.<\/p>\n<p><strong>Q: <\/strong>What are some of the lessons you\u2019ve learned from recent experiments, and challenges you\u2019ve had to overcome while designing these systems?<\/p>\n<p><strong>Carlone:<\/strong> Recently we did a big mapping experiment on the MIT campus, in which eight robots traversed up to 8 kilometers in total. The robots have no prior knowledge of the campus, and no GPS. Their main tasks are to estimate their own trajectory and build a map around it. You want the robots to understand the environment as humans do; humans not only understand the shape of obstacles, to get around them without hitting them, but also understand that an object is a chair, a desk, and so on. There\u2019s the semantics part.<\/p>\n<p>The interesting thing is that when the robots meet each other, they exchange information to improve their map of the environment. For instance, if robots connect, they can leverage information to correct their own trajectory. The challenge is that if you want to reach a consensus between robots, you don\u2019t have the bandwidth to exchange too much data. One of the key contributions of our 2022 paper is to deploy a distributed protocol, in which robots exchange limited information but can still agree on how the map looks. They don\u2019t send camera images back and forth but only exchange specific 3D coordinates and clues extracted from the sensor data. As they continue to exchange such data, they can form a consensus.<\/p>\n<p>Right now we are building color-coded 3D meshes or maps, in which the color contains some semantic information, like \u201cgreen\u201d corresponds to grass, and \u201cmagenta\u201d to a building. But as humans, we have a much more sophisticated understanding of reality, and we have a lot of prior knowledge about relationships between objects. For instance, if I was looking for a bed, I would go to the bedroom instead of exploring the entire house. If you start to understand the complex relationships between things, you can be much smarter about what the robot can do in the environment. We\u2019re trying to move from capturing just one layer of semantics, to a more hierarchical representation in which the robots understand rooms, buildings, and other concepts.<\/p>\n<p><strong>Q: <\/strong>What kinds of applications might Kimera and similar technologies lead to in the future?<\/p>\n<p><strong>How: <\/strong>Autonomous vehicle companies are doing a lot of mapping of the world and learning from the environments they\u2019re in. The holy grail would be if these vehicles could communicate with each other and share information, then they could improve models and maps that much quicker. The current solutions out there are individualized. If a truck pulls up next to you, you can\u2019t see in a certain direction. Could another vehicle provide a field of view that your vehicle otherwise doesn\u2019t have? This is a futuristic idea because it requires vehicles to communicate in new ways, and there are privacy issues to overcome. But if we could resolve those issues, you could imagine a significantly improved safety situation, where you have access to data from multiple perspectives, not only your field of view.<\/p>\n<p><strong>Carlone: <\/strong>These technologies will have a lot of applications. Earlier I mentioned search and rescue. Imagine that you want to explore a forest and look for survivors, or map buildings after an earthquake in a way that can help first responders access people who are trapped. Another setting where these technologies could be applied is in factories. Currently, robots that are deployed in factories are very rigid. They follow patterns on the floor, and are not really able to understand their surroundings. But if you\u2019re thinking about much more flexible factories in the future, robots will have to cooperate with humans and exist in a much less structured environment.<\/p>\n<\/div>\n<p><a href=\"https:\/\/news.mit.edu\/2023\/honing-robot-perception-mapping-0710\">Go to Source<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Author: Madeleine Turner | MIT Laboratory for Information and Decision Systems Walking to a friend\u2019s house or browsing the aisles of a grocery store might [&hellip;] <span class=\"read-more-link\"><a class=\"read-more\" href=\"https:\/\/www.aiproblog.com\/index.php\/2023\/07\/10\/3-questions-honing-robot-perception-and-mapping\/\">Read More<\/a><\/span><\/p>\n","protected":false},"author":1,"featured_media":471,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_bbp_topic_count":0,"_bbp_reply_count":0,"_bbp_total_topic_count":0,"_bbp_total_reply_count":0,"_bbp_voice_count":0,"_bbp_anonymous_reply_count":0,"_bbp_topic_count_hidden":0,"_bbp_reply_count_hidden":0,"_bbp_forum_subforum_count":0,"footnotes":""},"categories":[24],"tags":[],"_links":{"self":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/6632"}],"collection":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/comments?post=6632"}],"version-history":[{"count":0,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/6632\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media\/466"}],"wp:attachment":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media?parent=6632"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/categories?post=6632"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/tags?post=6632"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}