{"id":2934,"date":"2019-12-17T06:33:56","date_gmt":"2019-12-17T06:33:56","guid":{"rendered":"https:\/\/www.aiproblog.com\/index.php\/2019\/12\/17\/is-ai-about-to-hit-a-wall\/"},"modified":"2019-12-17T06:33:56","modified_gmt":"2019-12-17T06:33:56","slug":"is-ai-about-to-hit-a-wall","status":"publish","type":"post","link":"https:\/\/www.aiproblog.com\/index.php\/2019\/12\/17\/is-ai-about-to-hit-a-wall\/","title":{"rendered":"Is AI About to Hit a Wall?"},"content":{"rendered":"<p>Author: William Vorhies<\/p>\n<div>\n<p><strong><em>Summary:<\/em><\/strong> <em>\u00a0There have been several stories over the last several months around the theme that AI is about to hit a wall.\u00a0 That the rapid improvements we\u2019ve experienced and the benefits we\u2019ve accrued can\u2019t continue at the current pace.\u00a0 It\u2019s worth taking a look at these arguments to see if we should be adjusting our plans and expectations.<\/em><\/p>\n<p>\u00a0<\/p>\n<p><a href=\"https:\/\/storage.ning.com\/topology\/rest\/1.0\/file\/get\/3773143047?profile=original\" target=\"_blank\" rel=\"noopener noreferrer\"><img decoding=\"async\" src=\"https:\/\/storage.ning.com\/topology\/rest\/1.0\/file\/get\/3773143047?profile=RESIZE_710x\" width=\"250\" class=\"align-right\"><\/a>There have been several stories over the last several months around the theme that AI is about to hit a wall.\u00a0 That the rapid improvements we\u2019ve experienced and the benefits we\u2019ve accrued can\u2019t continue at the current pace.\u00a0 It\u2019s worth taking a look at these arguments to see if we should be adjusting our plans and expectations.<\/p>\n<p>These concerns center around two issues:<\/p>\n<ul>\n<li>The over concentration of investment in AI in just a few cities and how will that impact growth.<\/li>\n<li>The amount of compute required for our most advanced applications appears to be rising seven times faster than before implying some limit on the cost\/benefit to pursuing the most compute-hungry advances.<\/li>\n<\/ul>\n<p><span style=\"font-size: 12pt;\"><strong>Over Concentrated Investment<\/strong><\/span><\/p>\n<p>It\u2019s certainly true that investment in AI is extremely concentrated in just a few cities.\u00a0 This chart from the National Venture Capital Association shows that over 80% of investment went to just six metros.\u00a0 By the way, there\u2019s a sharp break in the distribution below this level with the next best metro receiving only 2.4% of total investment (Chicago).<\/p>\n<p>\u00a0<a href=\"https:\/\/storage.ning.com\/topology\/rest\/1.0\/file\/get\/3773143998?profile=original\" target=\"_blank\" rel=\"noopener noreferrer\"><img decoding=\"async\" src=\"https:\/\/storage.ning.com\/topology\/rest\/1.0\/file\/get\/3773143998?profile=RESIZE_710x\" width=\"500\" class=\"align-center\"><\/a><\/p>\n<p>So what\u2019s the concern?\u00a0 The folks who have written about this see this as an over concentration of talent that increases labor costs in these areas making some types on new AI investment uneconomic.<\/p>\n<p>It seems true that the concentration of AI talent in say San Francisco leads to the most new ideas in AI being generated there.\u00a0 But is this really a black hole with gravity so extreme that good ideas can\u2019t escape to be executed elsewhere?<\/p>\n<p>Worriers about this concentration see these good ideas escaping to overseas low cost locations.\u00a0 I\u2019m not so sure.\u00a0<\/p>\n<p>There is a pattern that the largest concentration of VC funded AI startups and established AI companies will draw in the most volume of new talent at the bottom.\u00a0 New graduates and young people will naturally gravitate to metros where there are the most opportunities.<\/p>\n<p>Personally I think this concern overlooks that as these metros become more dense with opportunity they also become massively more expensive to live in and that the quality of life drops off quickly (San Francisco, I\u2019m looking at you).<\/p>\n<p>As young professionals gain experience and also reach the life stage where they want to marry, buy homes, and start families they will naturally migrate to less expensive geographies.\u00a0 This may mean they will gravitate to the outskirts of these metros with lower cost of living where AI businesses will naturally follow to capture talent.\u00a0 It may also mean a willingness to relocate over longer distances to good quality of life locations like Austin, Seattle, Washington, Atlanta, or Miami where investment will follow talent.<\/p>\n<\/p>\n<p><span style=\"font-size: 12pt;\"><strong>Growth in Training Compute for the Most Advanced Applications is Now Growing 7X Faster than Before<\/strong><\/span><\/p>\n<p>This area of concern was first raised by the AI research lab OpenAI in 2018 and recently updated with new data.\u00a0 Their findings are that the amount of computational power needed to train the largest AI models has doubled every 3.4 months since 2012.<\/p>\n<p>They see history divided into two distinct eras with the new era beginning in 2012 now growing 7X faster than the previous era.<\/p>\n<p>\u00a0<a href=\"https:\/\/storage.ning.com\/topology\/rest\/1.0\/file\/get\/3773144931?profile=original\" target=\"_blank\" rel=\"noopener noreferrer\"><img decoding=\"async\" src=\"https:\/\/storage.ning.com\/topology\/rest\/1.0\/file\/get\/3773144931?profile=RESIZE_710x\" width=\"550\" class=\"align-center\"><\/a><\/p>\n<p>This phenomenon has led some leading industry experts including Jerome Pesenti, the VP of Artificial Intelligence at Facebook to observe <a href=\"https:\/\/www.wired.com\/story\/facebooks-ai-says-field-hit-wall\/\"><em><u>in a recent Wired article<\/u><\/em><\/a> that there may be cost\/benefit limits to what can be accomplished at the extreme limits of improving AI models.<\/p>\n<p>Pesenti says that while scaling up is a worthwhile solution to the most difficult challenges like deep fake image detection and AI moderation of fake news and online abuse that the rate of progress is not sustainable.<\/p>\n<p><em>\u201cIf you look at top experiments, each year the cost is going up 10-fold. Right now, an experiment might be in seven figures, but it\u2019s not going to go to nine or ten figures, it\u2019s not possible, nobody can afford that.\u00a0 It means that at some point we&#8217;re going to hit the wall. In many ways we already have.\u201d<\/em><\/p>\n<p>Other advancements that are at risk would be many of the <a href=\"https:\/\/www.datasciencecentral.com\/profiles\/blogs\/the-next-big-thing-in-ai-ml-is\"><em><u>\u2018next big things\u2019 in AI<\/u><\/em><\/a> including cutting edge applications that require increased look forwards or look backs to properly classify.\u00a0 These would include many navigation apps in next gen computer vision, and <a href=\"https:\/\/www.datasciencecentral.com\/profiles\/blogs\/contextually-intelligent-nlp-assistants-ai-s-next-big-technical-c\"><em><u>contextually intelligent NLP assistants<\/u><\/em><\/a> that can relate our schedules and actions to outside events.<\/p>\n<p>While I wouldn\u2019t want to have to do without these improvements, it\u2019s interesting what\u2019s been included in the OpenAI chart.\u00a0 For example, the most compute-hungry model is actually a reinforcement learning game playing app AlphaGoZero.\u00a0 Personally I\u2019d like to see research continue in Deep Learning RL models but the commercial applications are not as compelling as say BERT, the new NLP language encoder from Google.<\/p>\n<p>Another issue to keep in mind is that this is an issue that impacts incremental gains in existing capabilities.\u00a0 Pesenti is right to wonder aloud if it isn\u2019t time to apply cost\/benefit criteria to these.\u00a0 This so-called wall does not prevent us in any way from continuing the rapid application of existing AI\/ML capabilities in business and harvesting the benefits currently underway.<\/p>\n<p>In a sense this is the flying-car syndrome.\u00a0 We\u2019ve come so far so fast that if we don\u2019t get deep fake detection or fake news automatic moderation right away we\u2019ll wonder where\u2019s our flying car.<\/p>\n<p>It\u2019s also worth noting that lots of research is being conducted in how to make the training process more efficient.\u00a0 For example, just last week researchers at Rice University announced the results of their new MACH technique (Merged Average Classifiers via Hashing).\u00a0 <a href=\"https:\/\/news.rice.edu\/2019\/12\/09\/rice-amazon-report-breakthrough-in-distributed-deep-learning\/\"><em><u>They report<\/u><\/em><\/a> \u201ctraining times about 7-10 times faster, and \u2026memory footprints are 2-4 times smaller\u2019 than previous large-scale deep learning techniques.<\/p>\n<p>As for these walls, fortunately smart folks are working on both ends of this problem.\u00a0 We may need to bring some cost\/benefit discipline to our research but if the market need is there, my bet is that we\u2019ll find a way.<\/p>\n<p>\u00a0<\/p>\n<p><a href=\"https:\/\/www.datasciencecentral.com\/profiles\/blog\/list?user=0h5qapp2gbuf8\"><em><u>Other articles by Bill Vorhies<\/u><\/em><\/a><\/p>\n<p>\u00a0<\/p>\n<p>About the author:\u00a0 Bill is Contributing Editor for Data Science Central.\u00a0 Bill is also President &#038; Chief Data Scientist at Data-Magnum and has practiced as a data scientist since 2001.\u00a0 His articles have been read more than 2 million times.<\/p>\n<p>He can be reached at:<\/p>\n<p><a href=\"mailto:Bill@DataScienceCentral.com\">Bill@DataScienceCentral.com<\/a> <span>or<\/span> <a href=\"mailto:Bill@Data-Magnum.com\">Bill@Data-Magnum.com<\/a><\/p>\n<\/div>\n<p><a href=\"https:\/\/www.datasciencecentral.com\/xn\/detail\/6448529:BlogPost:915425\">Go to Source<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Author: William Vorhies Summary: \u00a0There have been several stories over the last several months around the theme that AI is about to hit a wall.\u00a0 [&hellip;] <span class=\"read-more-link\"><a class=\"read-more\" href=\"https:\/\/www.aiproblog.com\/index.php\/2019\/12\/17\/is-ai-about-to-hit-a-wall\/\">Read More<\/a><\/span><\/p>\n","protected":false},"author":1,"featured_media":474,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_bbp_topic_count":0,"_bbp_reply_count":0,"_bbp_total_topic_count":0,"_bbp_total_reply_count":0,"_bbp_voice_count":0,"_bbp_anonymous_reply_count":0,"_bbp_topic_count_hidden":0,"_bbp_reply_count_hidden":0,"_bbp_forum_subforum_count":0,"footnotes":""},"categories":[26],"tags":[],"_links":{"self":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/2934"}],"collection":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/comments?post=2934"}],"version-history":[{"count":0,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/2934\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media\/472"}],"wp:attachment":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media?parent=2934"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/categories?post=2934"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/tags?post=2934"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}