{"id":2579,"date":"2019-09-16T18:10:01","date_gmt":"2019-09-16T18:10:01","guid":{"rendered":"https:\/\/www.aiproblog.com\/index.php\/2019\/09\/16\/what-a-little-more-computing-power-can-do\/"},"modified":"2019-09-16T18:10:01","modified_gmt":"2019-09-16T18:10:01","slug":"what-a-little-more-computing-power-can-do","status":"publish","type":"post","link":"https:\/\/www.aiproblog.com\/index.php\/2019\/09\/16\/what-a-little-more-computing-power-can-do\/","title":{"rendered":"What a little more computing power can do"},"content":{"rendered":"<p>Author: Kim Martineau | MIT Quest for Intelligence<\/p>\n<div>\n<p>Neural networks have given researchers a powerful tool for looking into the future and making predictions. But one drawback is their insatiable need for data and computing power (&#8220;compute&#8221;) to process all that information. At MIT, demand for compute is estimated to be five times greater than what the Institute can offer. To help ease the crunch, industry has stepped in. An $11.6 million supercomputer recently\u00a0<a href=\"http:\/\/news.mit.edu\/2019\/ibm-gives-lift-artificial-intelligence-computing-mit-0826\">donated by IBM<\/a>\u00a0comes online this fall, and in the past year, both IBM and Google have provided cloud credits\u00a0to MIT Quest for Intelligence for distribution across campus. Four projects made possible by IBM and Google cloud donations are highlighted below.<\/p>\n<p><strong>Smaller, faster, smarter neural networks<\/strong><\/p>\n<p>To recognize a cat in a picture, a deep learning model may need to see millions of photos before its artificial neurons \u201clearn\u201d to identify a cat. The process is computationally intensive and carries a steep\u00a0<a href=\"https:\/\/www.technologyreview.com\/s\/613630\/training-a-single-ai-model-can-emit-as-much-carbon-as-five-cars-in-their-lifetimes\/\">environmental cost<\/a>,\u00a0as new research attempting to measure artificial intelligence&#8217;s (AI\u2019s) carbon footprint has highlighted.\u00a0<\/p>\n<p>But there may be a more efficient way. New MIT research shows that models only a fraction of the size are needed. \u201cWhen you train a big network there\u2019s a small one that could have done everything,\u201d says\u00a0<a href=\"http:\/\/www.jfrankle.com\/\">Jonathan Frankle<\/a>, a graduate student in MIT\u2019s\u00a0<a href=\"https:\/\/www.eecs.mit.edu\/\">Department of Electrical Engineering and Computer Science<\/a> (EECS).<\/p>\n<p>With study co-author and EECS Professor\u00a0<a href=\"https:\/\/people.csail.mit.edu\/mcarbin\/\">Michael Carbin<\/a>, Frankle\u00a0<a href=\"http:\/\/news.mit.edu\/2019\/smarter-training-neural-networks-0506\">estimates<\/a>\u00a0that a neural network could get by with on-tenth the number of connections if the right subnetwork is found at the outset. Normally, neural networks are trimmed after the training process, with irrelevant connections removed then.\u00a0Why not train the small model to begin with, Frankle wondered?<\/p>\n<p>Experimenting with a two-neuron network on his laptop, Frankle got encouraging results and moved to larger image-datasets like MNIST and CIFAR-10, borrowing GPUs where he could. Finally, through IBM Cloud, he secured enough compute power to train a real ResNet model. \u201cEverything I\u2019d done previously was toy experiments,\u201d he says. \u201cI was finally able to run dozens of different settings to make sure I could make the claims in our paper.\u201d<\/p>\n<p>Frankle spoke from Facebook\u2019s offices, where he worked for the summer to explore ideas raised by his\u00a0<a href=\"https:\/\/arxiv.org\/pdf\/1803.03635.pdf\">Lottery Ticket Hypothesis<\/a>\u00a0paper, one of two picked for a best paper award at this year\u2019s International Conference on Learning Representations. Potential applications for the work go beyond image classification, Frankle says, and include reinforcement learning and natural language processing models. Already, researchers at\u00a0<a href=\"https:\/\/arxiv.org\/pdf\/1906.02773.pdf\">Facebook AI Research<\/a>,\u00a0<a href=\"https:\/\/arxiv.org\/pdf\/1905.07785.pdf\">Princeton University<\/a>, and\u00a0<a href=\"https:\/\/arxiv.org\/pdf\/1905.01067.pdf\">Uber<\/a>\u00a0have published follow-on studies.\u00a0<\/p>\n<p>\u201cWhat I love about neural networks is we haven\u2019t even laid the foundation yet,\u201d says Frankle, who recently shifted from studying cryptography and tech policy to AI. \u201cWe really don\u2019t understand how it learns, where it\u2019s good and where it fails. This is physics 1,000 years before Newton.\u201d<\/p>\n<p><strong>Distinguishing fact from fake news<\/strong><\/p>\n<p>Networking platforms like Facebook and Twitter have made it easier than ever to find quality news. But too often, real news is drowned out by misleading or outright false information posted online. Confusion over a recent video of U.S. House Speaker Nancy Pelosi doctored to make her sound drunk is just the latest example of the threat misinformation and fake news pose to democracy.\u00a0<\/p>\n<p>\u201cYou can put just about anything up on the internet now, and some people will believe it,\u201d says\u00a0<a href=\"https:\/\/moinnadeem.com\/\" target=\"_blank\" rel=\"noopener noreferrer\">Moin Nadeem<\/a>, a senior and EECS major at MIT.<\/p>\n<p>If technology helped create the problem, it can also help fix it. That was Nadeem\u2019s reason for picking a\u00a0<a href=\"https:\/\/www.youtube.com\/watch?v=U5PXVZ_kjzI\">superUROP<\/a>\u00a0project focused on building an automated system to fight fake and misleading news. Working in the lab of James Glass, a researcher at MIT\u2019s\u00a0<a href=\"https:\/\/www.csail.mit.edu\/person\/jim-glass\" target=\"_blank\" rel=\"noopener noreferrer\">Computer Science and Artificial Intelligence Laboratory<\/a>, and supervised by Mitra Mohtarami, Nadeem helped train a language model to fact-check claims by searching through Wikipedia and three types of news sources rated by journalists as high-quality, mixed-quality or low-quality.<\/p>\n<p>To verify a claim, the model measures how closely the sources agree, with higher agreement scores indicating the claim is likely true. A high disagreement score for a claim like, \u201cISIS infiltrates the United States,\u201d is a strong indicator of fake news. One drawback of this method, he says, is that the model doesn\u2019t identify the independent truth so much as describe what most people think is true.<\/p>\n<p>With the help of Google Cloud Platform, Nadeem ran experiments and built an interactive website that lets users instantly assess the accuracy of a claim. He and his co-authors presented their\u00a0<a href=\"https:\/\/arxiv.org\/abs\/1906.04164\" target=\"_blank\" rel=\"noopener noreferrer\">results<\/a>\u00a0at the North American Association of Computational Linguistics (NAACL) conference in June and are continuing to expand on the work.<\/p>\n<p>\u201cThe saying used to be that seeing is believing,\u201d says Nadeem, in\u00a0<a href=\"https:\/\/www.youtube.com\/watch?v=tliyKmxEsw8\">this video<\/a>\u00a0about his work. \u201cBut we\u2019re entering a world where that isn\u2019t true. If people can\u2019t trust their eyes and ears it becomes a question of what\u00a0can<em>\u00a0<\/em>we trust?\u201d<\/p>\n<p><strong>Visualizing a warming climate<\/strong><\/p>\n<p>From rising seas to increased droughts, the effects of climate change are already being felt. A few decades from now, the world will be a warmer, drier, and more unpredictable place.\u00a0<a href=\"https:\/\/www.linkedin.com\/in\/brandonl0l\/\">Brandon Leshchinskiy<\/a>, a graduate student in MIT\u2019s\u00a0<a href=\"https:\/\/aeroastro.mit.edu\/\">Department of Aeronautics and Astronautics<\/a>\u00a0(AeroAstro), is experimenting with generative adversarial networks, or GANs, to imagine what Earth will look like then.\u00a0<\/p>\n<p>GANs produce hyper-realistic imagery by pitting one neural network against another. The first network learns the underlying structure of a set of images and tries to reproduce them, while the second decides which images look implausible and tells the first network to try again.<\/p>\n<p>Inspired by researchers who used GANs to\u00a0<a href=\"https:\/\/www.technologyreview.com\/f\/613547\/ai-can-show-us-the-ravages-of-climate-change\/\">visualize<\/a>\u00a0sea-level rise projections from street-view images, Leshchinskiy wanted to see if satellite imagery could similarly personalize climate projections. With his advisor, AeroAstro Professor\u00a0<a href=\"http:\/\/web.mit.edu\/aeroastro\/www\/people\/dnewman\/bio.html\">Dava Newman<\/a>, Leshchinskiy is currently using free IBM Cloud credits to train a pair of GANs on images of the eastern U.S. coastline with their corresponding elevation points. The goal is to visualize how sea-level rise projections for 2050 will redraw the coastline. If the project works, Leshinskiy hopes to use other NASA datasets to imagine future ocean acidification and changes in phytoplankton abundance.\u00a0<\/p>\n<p>\u201cWe\u2019re past the point of mitigation,\u201d he says. \u201cVisualizing what the world will look like three decades from now can help us adapt to climate change.\u201d<\/p>\n<p><strong>Identifying athletes from a few gestures<\/strong><\/p>\n<p>A few moves on the field or court are enough for a computer vision model to identify individual athletes. That\u2019s according to preliminary research by a team led by\u00a0<a href=\"https:\/\/www.linkedin.com\/in\/katherinevgallagher\/\">Katherine Gallagher<\/a>, a researcher at MIT Quest for Intelligence.<\/p>\n<p>The team trained computer vision models\u00a0on video recordings of tennis matches and soccer and basketball games\u00a0and found that the models could recognize individual players in just a few frames from key points on their body providing a rough outline of their skeleton.\u00a0<\/p>\n<p>The team used a Google Cloud API to process the video data, and compared their models&#8217; performance against\u00a0models trained on Google Cloud&#8217;s AI platform.\u00a0\u201cThis pose information is so distinctive that our models\u00a0can identify players with accuracy almost as good as models provided with\u00a0much\u00a0more information,\u00a0like hair color and clothing,\u201d she says.\u00a0<\/p>\n<p>Their results are relevant for automated player identification in sports analytics systems, and they could provide a basis for further research on inferring player fatigue to anticipate when players should be swapped out. Automated pose detection could also help athletes refine their technique by allowing them to isolate the precise moves associated with a golfer\u2019s expert drive or a tennis player\u2019s winning swing.<\/p>\n<\/div>\n<p><a href=\"http:\/\/news.mit.edu\/2019\/what-extra-computing-power-can-do-0916\">Go to Source<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Author: Kim Martineau | MIT Quest for Intelligence Neural networks have given researchers a powerful tool for looking into the future and making predictions. But [&hellip;] <span class=\"read-more-link\"><a class=\"read-more\" href=\"https:\/\/www.aiproblog.com\/index.php\/2019\/09\/16\/what-a-little-more-computing-power-can-do\/\">Read More<\/a><\/span><\/p>\n","protected":false},"author":1,"featured_media":471,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_bbp_topic_count":0,"_bbp_reply_count":0,"_bbp_total_topic_count":0,"_bbp_total_reply_count":0,"_bbp_voice_count":0,"_bbp_anonymous_reply_count":0,"_bbp_topic_count_hidden":0,"_bbp_reply_count_hidden":0,"_bbp_forum_subforum_count":0,"footnotes":""},"categories":[24],"tags":[],"_links":{"self":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/2579"}],"collection":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/comments?post=2579"}],"version-history":[{"count":0,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/2579\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media\/471"}],"wp:attachment":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media?parent=2579"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/categories?post=2579"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/tags?post=2579"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}