{"id":7411,"date":"2024-06-18T04:00:00","date_gmt":"2024-06-18T04:00:00","guid":{"rendered":"https:\/\/www.aiproblog.com\/index.php\/2024\/06\/18\/researchers-leverage-shadows-to-model-3d-scenes-including-objects-blocked-from-view\/"},"modified":"2024-06-18T04:00:00","modified_gmt":"2024-06-18T04:00:00","slug":"researchers-leverage-shadows-to-model-3d-scenes-including-objects-blocked-from-view","status":"publish","type":"post","link":"https:\/\/www.aiproblog.com\/index.php\/2024\/06\/18\/researchers-leverage-shadows-to-model-3d-scenes-including-objects-blocked-from-view\/","title":{"rendered":"Researchers leverage shadows to model 3D scenes, including objects blocked from view"},"content":{"rendered":"<p>Author: Adam Zewe | MIT News<\/p>\n<div>\n<p>Imagine driving through a tunnel in an autonomous vehicle, but unbeknownst to you, a crash has stopped traffic up ahead. Normally, you\u2019d need to rely on the car in front of you to know you should start braking. But what if your vehicle could see around the car ahead and apply the brakes even sooner?<\/p>\n<p>Researchers from MIT and Meta have developed a computer vision technique that could someday enable an autonomous vehicle to do just that.<\/p>\n<p>They have introduced a method that creates physically accurate, 3D models of an entire scene, including areas blocked from view, using images from a single camera position. Their technique uses shadows to determine what lies in obstructed portions of the scene.<\/p>\n<p>They call their approach PlatoNeRF, based on Plato\u2019s allegory of the cave, a passage from the Greek philosopher\u2019s \u201cRepublic\u201d<em>\u00a0<\/em>in which prisoners chained in a cave discern the reality of the outside world based on shadows cast on the cave wall.<\/p>\n<p>By combining lidar (light detection and ranging) technology with machine learning, PlatoNeRF can generate more accurate reconstructions of 3D geometry than some existing AI techniques. Additionally, PlatoNeRF is better at smoothly reconstructing scenes where shadows are hard to see, such as those with high ambient light or dark backgrounds.<\/p>\n<p>In addition to improving the safety of autonomous vehicles, PlatoNeRF could make AR\/VR headsets more efficient by enabling a user to model the geometry of a room without the need to walk around taking measurements. It could also help warehouse robots find items in cluttered environments faster.<\/p>\n<p>\u201cOur key idea was taking these two things that have been done in different disciplines before and pulling them together \u2014 multibounce lidar and machine learning. It turns out that when you bring these two together, that is when you find a lot of new opportunities to explore and get the best of both worlds,\u201d says Tzofi Klinghoffer, an MIT graduate student in media arts and sciences, affiliate of the MIT Media Lab, and lead author of a <a href=\"https:\/\/openaccess.thecvf.com\/content\/CVPR2024\/html\/Klinghoffer_PlatoNeRF_3D_Reconstruction_in_Platos_Cave_via_Single-View_Two-Bounce_Lidar_CVPR_2024_paper.html\" target=\"_blank\" rel=\"noopener\">paper on PlatoNeRF<\/a>.<\/p>\n<p>Klinghoffer wrote the paper with his advisor, Ramesh Raskar, associate professor of media arts and sciences and leader of the Camera Culture Group at MIT; senior author Rakesh Ranjan, a director of AI research at Meta Reality Labs; as well as Siddharth Somasundaram at MIT, and Xiaoyu Xiang, Yuchen Fan, and Christian Richardt at Meta. The research will be presented at the Conference on Computer Vision and Pattern Recognition.<\/p>\n<p><strong>Shedding light on the problem<\/strong><\/p>\n<p>Reconstructing a full 3D scene from one camera viewpoint is a complex problem.<\/p>\n<p>Some machine-learning approaches employ generative AI models that try to guess what lies in the occluded regions, but these models can hallucinate objects that aren\u2019t really there. Other approaches attempt to infer the shapes of hidden objects using shadows in a color image, but these methods can struggle when shadows are hard to see.<\/p>\n<p>For PlatoNeRF, the MIT researchers built off these approaches using a new sensing modality called single-photon lidar. Lidars map a 3D scene by emitting pulses of light and measuring the time it takes that light to bounce back to the sensor. Because single-photon lidars can detect individual photons, they provide higher-resolution data.<\/p>\n<p>The researchers use a single-photon lidar to illuminate a target point in the scene. Some light bounces off that point and returns directly to the sensor. However, most of the light scatters and bounces off other objects before returning to the sensor. PlatoNeRF relies on these second bounces of light.<\/p>\n<p>By calculating how long it takes light to bounce twice and then return to the lidar sensor, PlatoNeRF captures additional information about the scene, including depth. The second bounce of light also contains information about shadows.<\/p>\n<p>The system traces the secondary rays of light \u2014 those that bounce off the target point to other points in the scene \u2014 to determine which points lie in shadow (due to an absence of light). Based on the location of these shadows, PlatoNeRF can infer the geometry of hidden objects.<\/p>\n<p>The lidar sequentially illuminates 16 points, capturing multiple images that are used to reconstruct the entire 3D scene.<\/p>\n<p>\u201cEvery time we illuminate a point in the scene, we are creating new shadows. Because we have all these different illumination sources, we have a lot of light rays shooting around, so we are carving out the region that is occluded and lies beyond the visible eye,\u201d Klinghoffer says.<\/p>\n<p><strong>A winning combination<\/strong><\/p>\n<p>Key to PlatoNeRF is the combination of multibounce lidar with a special type of machine-learning model known as a neural radiance field (NeRF). A NeRF encodes the geometry of a scene into the weights of a neural network, which gives the model a strong ability to interpolate, or estimate, novel views of a scene.<\/p>\n<p>This ability to interpolate also leads to highly accurate scene reconstructions when combined with multibounce lidar, Klinghoffer says.<\/p>\n<p>\u201cThe biggest challenge was figuring out how to combine these two things. We really had to think about the physics of how light is transporting with multibounce lidar and how to model that with machine learning,\u201d he says.<\/p>\n<p>They compared PlatoNeRF to two common alternative methods, one that only uses lidar and the other that only uses a NeRF with a color image.<\/p>\n<p>They found that their method was able to outperform both techniques, especially when the lidar sensor had lower resolution. This would make their approach more practical to deploy in the real world, where lower resolution sensors are common in commercial devices.<\/p>\n<p>\u201cAbout 15 years ago, our group invented the first camera to \u2018see\u2019 around corners, that works by exploiting multiple bounces of light, or \u2018echoes of light.\u2019 Those techniques used special lasers and sensors, and used three bounces of light. Since then, lidar technology has become more mainstream, that led to our research on cameras that can see through fog. This new work uses only two bounces of light, which means the signal to noise ratio is very high, and 3D reconstruction quality is impressive,\u201d Raskar says.<\/p>\n<p>In the future, the researchers want to try tracking more than two bounces of light to see how that could improve scene reconstructions. In addition, they are interested in applying more deep learning techniques and combining PlatoNeRF with color image measurements to capture texture information.<\/p>\n<p>\u201cWhile camera images of shadows have long been studied as a means to 3D reconstruction, this work revisits the problem in the context of lidar, demonstrating significant improvements in the accuracy of reconstructed hidden geometry. The work shows how clever algorithms can enable extraordinary capabilities when combined with ordinary sensors \u2014 including the lidar systems that many of us now carry in our pocket,\u201d says David Lindell, an assistant professor in the Department of Computer Science at the University of Toronto, who was not involved with this work.<\/p>\n<\/div>\n<p><a href=\"https:\/\/news.mit.edu\/2024\/researchers-leverage-shadows-model-3d-scenes-blocked-objects-0618\">Go to Source<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Author: Adam Zewe | MIT News Imagine driving through a tunnel in an autonomous vehicle, but unbeknownst to you, a crash has stopped traffic up [&hellip;] <span class=\"read-more-link\"><a class=\"read-more\" href=\"https:\/\/www.aiproblog.com\/index.php\/2024\/06\/18\/researchers-leverage-shadows-to-model-3d-scenes-including-objects-blocked-from-view\/\">Read More<\/a><\/span><\/p>\n","protected":false},"author":1,"featured_media":469,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_bbp_topic_count":0,"_bbp_reply_count":0,"_bbp_total_topic_count":0,"_bbp_total_reply_count":0,"_bbp_voice_count":0,"_bbp_anonymous_reply_count":0,"_bbp_topic_count_hidden":0,"_bbp_reply_count_hidden":0,"_bbp_forum_subforum_count":0,"footnotes":""},"categories":[24],"tags":[],"_links":{"self":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/7411"}],"collection":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/comments?post=7411"}],"version-history":[{"count":0,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/7411\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media\/474"}],"wp:attachment":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media?parent=7411"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/categories?post=7411"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/tags?post=7411"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}