{"id":7723,"date":"2024-11-05T05:00:00","date_gmt":"2024-11-05T05:00:00","guid":{"rendered":"http:\/\/54.175.255.16\/index.php\/2024\/11\/05\/despite-its-impressive-output-generative-ai-doesnt-have-a-coherent-understanding-of-the-world\/"},"modified":"2024-11-05T05:00:00","modified_gmt":"2024-11-05T05:00:00","slug":"despite-its-impressive-output-generative-ai-doesnt-have-a-coherent-understanding-of-the-world","status":"publish","type":"post","link":"https:\/\/www.aiproblog.com\/index.php\/2024\/11\/05\/despite-its-impressive-output-generative-ai-doesnt-have-a-coherent-understanding-of-the-world\/","title":{"rendered":"Despite its impressive output, generative AI doesn\u2019t have a coherent understanding of the world"},"content":{"rendered":"<p>Author: Adam Zewe | MIT News<\/p>\n<div>\n<p>Large language models can do impressive things, like write poetry or generate viable computer programs, even though these models are trained to predict words that come next in a piece of text.<\/p>\n<p>Such surprising capabilities can make it seem like the models are implicitly learning some general truths about the world.<\/p>\n<p>But that isn\u2019t necessarily the case, according to a new study. The researchers found that a popular type of\u00a0<a href=\"https:\/\/news.mit.edu\/2023\/explained-generative-ai-1109\" target=\"_blank\" rel=\"noopener\">generative AI model<\/a> can provide turn-by-turn driving directions in New York City with near-perfect accuracy \u2014 without having formed an accurate internal map of the city.<\/p>\n<p>Despite the model\u2019s uncanny ability to navigate effectively, when the researchers closed some streets and added detours, its performance plummeted.<\/p>\n<p>When they dug deeper, the researchers found that the New York maps the model implicitly generated had many nonexistent streets curving between the grid and connecting far away intersections.<\/p>\n<p>This could have serious implications for generative AI models deployed in the real world, since a model that seems to be performing well in one context might break down if the task or environment slightly changes.<\/p>\n<p>\u201cOne hope is that, because LLMs can accomplish all these amazing things in language, maybe we could use these same tools in other parts of science, as well. But the question of whether LLMs are learning coherent world models is very important if we want to use these techniques to make new discoveries,\u201d says senior author Ashesh Rambachan, assistant professor of economics and a principal investigator in the MIT Laboratory for Information and Decision Systems (LIDS).<\/p>\n<p>Rambachan is joined on a <a href=\"https:\/\/arxiv.org\/pdf\/2406.03689\" target=\"_blank\" rel=\"noopener\">paper about the work<\/a> by lead author Keyon Vafa, a postdoc at Harvard University; Justin Y. Chen, an electrical engineering and computer science (EECS) graduate student at MIT; Jon Kleinberg, Tisch University Professor of Computer Science and Information Science at Cornell University; and Sendhil Mullainathan, an MIT professor in the departments of EECS and of Economics, and a member of LIDS. The research will be presented at the Conference on Neural Information Processing Systems.<\/p>\n<p><strong>New metrics<\/strong><\/p>\n<p>The researchers focused on a type of generative AI model known as a transformer, which forms the backbone of LLMs like GPT-4. Transformers are trained on a massive amount of language-based data to predict the next token in a sequence, such as the next word in a sentence.<\/p>\n<p>But if scientists want to determine whether an LLM has formed an accurate model of the world, measuring the accuracy of its predictions doesn\u2019t go far enough, the researchers say.<\/p>\n<p>For example, they found that a transformer can predict valid moves in a game of Connect 4 nearly every time without understanding any of the rules.<\/p>\n<p>So, the team developed two new metrics that can test a transformer\u2019s world model. The researchers focused their evaluations on a class of problems called deterministic finite automations, or DFAs.\u00a0<\/p>\n<p>A DFA is a problem with a sequence of states, like intersections one must traverse to reach a destination, and a concrete way of describing the rules one must follow along the way.<\/p>\n<p>They chose two problems to formulate as DFAs: navigating on streets in New York City and playing the board game Othello.<\/p>\n<p>\u201cWe needed test beds where we know what the world model is. Now, we can rigorously think about what it means to recover that world model,\u201d Vafa explains.<\/p>\n<p>The first metric they developed, called sequence distinction, says a model has formed a coherent world model it if sees two different states, like two different Othello boards, and recognizes how they are different. Sequences, that is, ordered lists of data points, are what transformers use to generate outputs.<\/p>\n<p>The second metric, called sequence compression, says a transformer with a coherent world model should know that two identical states, like two identical Othello boards, have the same sequence of possible next steps.<\/p>\n<p>They used these metrics to test two common classes of transformers, one which is trained on data generated from randomly produced sequences and the other on data generated by following strategies.<\/p>\n<p><strong>Incoherent world models<\/strong><\/p>\n<p>Surprisingly, the researchers found that transformers which made choices randomly formed more accurate world models, perhaps because they saw a wider variety of potential next steps during training.\u00a0<\/p>\n<p>\u201cIn Othello, if you see two random computers playing rather than championship players, in theory you\u2019d see the full set of possible moves, even the bad moves championship players wouldn\u2019t make,\u201d Vafa explains.<\/p>\n<p>Even though the transformers generated accurate directions and valid Othello moves in nearly every instance, the two metrics revealed that only one generated a coherent world model for Othello moves, and none performed well at forming coherent world models in the wayfinding example.<\/p>\n<p>The researchers demonstrated the implications of this by adding detours to the map of New York City, which caused all the navigation models to fail.<\/p>\n<p>\u201cI was surprised by how quickly the performance deteriorated as soon as we added a detour. If we close just 1 percent of the possible streets, accuracy immediately plummets from nearly 100 percent to just 67 percent,\u201d Vafa says.<\/p>\n<p>When they recovered the city maps the models generated, they looked like an imagined New York City with hundreds of streets crisscrossing overlaid on top of the grid. The maps often contained random flyovers above other streets or multiple streets with impossible orientations.<\/p>\n<p>These results show that transformers can perform surprisingly well at certain tasks without understanding the rules. If scientists want to build LLMs that can capture accurate world models, they need to take a different approach, the researchers say.<\/p>\n<p>\u201cOften, we see these models do impressive things and think they must have understood something about the world. I hope we can convince people that this is a question to think very carefully about, and we don\u2019t have to rely on our own intuitions to answer it,\u201d says Rambachan.<\/p>\n<p>In the future, the researchers want to tackle a more diverse set of problems, such as those where some rules are only partially known. They also want to apply their evaluation metrics to real-world, scientific problems.<\/p>\n<p>This work is funded, in part, by the Harvard Data Science Initiative, a National Science Foundation Graduate Research Fellowship, a Vannevar Bush Faculty Fellowship, a Simons Collaboration grant, and a grant from the MacArthur Foundation.<\/p>\n<\/div>\n<p><a href=\"https:\/\/news.mit.edu\/2024\/generative-ai-lacks-coherent-world-understanding-1105\">Go to Source<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Author: Adam Zewe | MIT News Large language models can do impressive things, like write poetry or generate viable computer programs, even though these models [&hellip;] <span class=\"read-more-link\"><a class=\"read-more\" href=\"https:\/\/www.aiproblog.com\/index.php\/2024\/11\/05\/despite-its-impressive-output-generative-ai-doesnt-have-a-coherent-understanding-of-the-world\/\">Read More<\/a><\/span><\/p>\n","protected":false},"author":1,"featured_media":474,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_bbp_topic_count":0,"_bbp_reply_count":0,"_bbp_total_topic_count":0,"_bbp_total_reply_count":0,"_bbp_voice_count":0,"_bbp_anonymous_reply_count":0,"_bbp_topic_count_hidden":0,"_bbp_reply_count_hidden":0,"_bbp_forum_subforum_count":0,"footnotes":""},"categories":[24],"tags":[],"_links":{"self":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/7723"}],"collection":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/comments?post=7723"}],"version-history":[{"count":0,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/7723\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media\/468"}],"wp:attachment":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media?parent=7723"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/categories?post=7723"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/tags?post=7723"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}