{"id":1136,"date":"2018-10-09T06:37:35","date_gmt":"2018-10-09T06:37:35","guid":{"rendered":"https:\/\/www.aiproblog.com\/index.php\/2018\/10\/09\/automated-deep-learning-so-simple-anyone-can-do-it\/"},"modified":"2018-10-09T06:37:35","modified_gmt":"2018-10-09T06:37:35","slug":"automated-deep-learning-so-simple-anyone-can-do-it","status":"publish","type":"post","link":"https:\/\/www.aiproblog.com\/index.php\/2018\/10\/09\/automated-deep-learning-so-simple-anyone-can-do-it\/","title":{"rendered":"Automated Deep Learning \u2013 So Simple Anyone Can Do It"},"content":{"rendered":"<p>Author: William Vorhies<\/p>\n<div>\n<p><strong><em>Summary:<\/em><\/strong><em>\u00a0 There are several things holding back our use of deep learning methods and chief among them is that they are complicated and hard.\u00a0 Now there are three platforms that offer Automated Deep Learning (ADL) so simple that almost anyone can do it.<\/em><\/p>\n<p>\u00a0<\/p>\n<p><a href=\"http:\/\/api.ning.com\/files\/ds5VTXUtVmvlM4LI34YauH6cd-ytZVRraQjb*Tdy-xJatSZGw6ugSsJbu94kbk-AZ9ICGFqMWPUbD2hDDp8mi-TuCjmUrUDc\/HomerSimpsonComputer.png\" target=\"_self\"><img decoding=\"async\" src=\"http:\/\/api.ning.com\/files\/ds5VTXUtVmvlM4LI34YauH6cd-ytZVRraQjb*Tdy-xJatSZGw6ugSsJbu94kbk-AZ9ICGFqMWPUbD2hDDp8mi-TuCjmUrUDc\/HomerSimpsonComputer.png?width=350\" width=\"350\" class=\"align-right\"><\/a>There are several things holding back our use of deep learning methods and chief among them is that they are complicated and hard.\u00a0<\/p>\n<p>A small percentage of our data science community has chosen the path of learning these new techniques, but it\u2019s a major departure both in problem type and technique from the predictive and prescriptive modeling that makes up 90% of what we get paid to do.<\/p>\n<p>Artificial intelligence, at least in the true sense of image, video, text, and speech recognition and processing is on everyone\u2019s lips but it\u2019s still hard to find a data scientist qualified to execute your project.<\/p>\n<p>Actually when I list image, video, text, and speech applications I\u2019m selling deep learning a little short.\u00a0 While these are the best known and perhaps most obvious applications, deep neural nets (DNNs) are also proving excellent at forecasting time series data, and also in complex traditional consumer propensity problems.<\/p>\n<p>Last December as I was listing <a href=\"https:\/\/www.datasciencecentral.com\/profiles\/blogs\/6-predictions-about-data-science-machine-learning-and-ai-for-2018\"><em><u>my predictions for 2018<\/u><\/em><\/a>, I noted that Gartner had said that during 2018, DNNs would become a standard component in the toolbox of 80% of data scientists.\u00a0 My prediction was that while the first provider to accomplish this level of simplicity would certainly be richly rewarded, no way was it going to be 2018.\u00a0 It seems I was wrong.<\/p>\n<p>Here we are and it\u2019s only April and I\u2019ve recently been introduced to three different platforms that have the goal of making deep learning so easy, anyone (well at least any data scientist) can do it.<\/p>\n<\/p>\n<p><span style=\"font-size: 12pt;\"><strong>Minimum Requirements<\/strong><\/span><\/p>\n<p>All of the majors and several smaller companies offer greatly simplified tools for executing CNNs or RNN\/LSTMs, but these still require experimental hand tuning of the layer types and number, connectivity, nodes, and all the other hyperparameters that so often defeat initial success.\u00a0<\/p>\n<p>To be part of this group you need a truly one-click application that allows the average data scientists or even developer to build a successful image or text classifier.<\/p>\n<p>The quickest route to this goal is by transfer learning.\u00a0 In DL, transfer learning means taking a previously built successful, large, complex CNN or RNN\/LSTM model and using a new more limited data set to train against it.\u00a0<\/p>\n<p>Basically transfer learning, most used in image classification, summarizes the more complex model into fewer or previously trained categories.\u00a0 Transfer learning can\u2019t create classifications that weren\u2019t in the original model, but it can learn to create subsets or summary categories of what\u2019s there.<\/p>\n<p>The advantage is that the hyperparameter tuning has already been done so you know the model will train.\u00a0 More importantly, you can build a successful transfer model with just a few hundred labeled images in less than an hour.<\/p>\n<p>The real holy grail of AutoDL however, is fully automated hyperparameter tuning, not transfer learning.\u00a0 As you\u2019ll read below, some are on track, and others claim to already have succeeded.<\/p>\n<\/p>\n<p><span style=\"font-size: 12pt;\"><strong>Microsoft CustomVision.AI<\/strong><\/span><\/p>\n<p>Late in 2017 MS introduced a series of greatly simplified DL capabilities covering the full range of image, video, text, and speech under the banner of the Microsoft Cognitive Services.\u00a0 In January they introduced their fully automated platform, Microsoft Custom Vision Services (<a href=\"https:\/\/www.customvision.ai\/\">https:\/\/www.customvision.ai\/<\/a><span>).<\/span><\/p>\n<p><span>The platform is limited to image classifiers and promises to allow users to create robust CNN transfer models based on only a few images capitalizing on MS\u2019s huge existing library of large, complex, multi-image classifiers.<\/span><\/p>\n<p>Using the platform is extremely simple.\u00a0 You drag and drop your images onto the platform and press go.\u00a0 You\u2019ll need at least a pay-as-you-go Azure account and basic tech support runs $29\/mo.\u00a0 It\u2019s not clear how long the models take to train but since it\u2019s transfer learning it should be quite fast and therefore, we\u2019re guessing, inexpensive (but not free).<\/p>\n<p>During project setup you\u2019ll be asked to identify a general domain from which your image set will transfer learn and these currently are:<\/p>\n<ol>\n<li>General<\/li>\n<li>Food<\/li>\n<li>Landmarks<\/li>\n<li>Retail<\/li>\n<li>Adult<\/li>\n<li>General (compact)<\/li>\n<li>Landmarks (compact)<\/li>\n<li>Retail (compact)<\/li>\n<\/ol>\n<p>While all these models will run from a restful API once trained, the last three categories (marked \u2018compact\u2019) can be exported to run off line on any iOS or Android edge device.\u00a0 Export is to the CoreML format for iOS 11 and to the TensorFlow format for Android. \u00a0This should entice a variety of app developers who may not be data scientist to add instant image classification to their device.<\/p>\n<p>You can bet MS will be rolling out more complex features as fast as possible.<\/p>\n<\/p>\n<p><span style=\"font-size: 12pt;\"><strong>Google Cloud AutoML<\/strong><\/span><\/p>\n<p>Also in January, Google announced its similar entry <a href=\"https:\/\/cloud.google.com\/automl\/\"><em><u>Cloud AutoML<\/u><\/em><\/a>.\u00a0 The platform is in alpha and requires an invitation to participate.<\/p>\n<p>Like Microsoft, the service utilizes transfer learning from Google\u2019s own prebuilt complex CNN classifiers.\u00a0 They recommend at least 100 images per label for transfer learning.\u00a0<\/p>\n<p><a href=\"http:\/\/api.ning.com\/files\/ds5VTXUtVmt-TKxoiTaWIyo9We69u24ZQ8hwY7aX7AXoljpnxpEZYR2gwPyCMhpNkp7Zoz7y8wx4ec3Q6s*X1s3OTcuv*WOA\/GoogleAutoML1.png\" target=\"_self\"><img decoding=\"async\" src=\"http:\/\/api.ning.com\/files\/ds5VTXUtVmt-TKxoiTaWIyo9We69u24ZQ8hwY7aX7AXoljpnxpEZYR2gwPyCMhpNkp7Zoz7y8wx4ec3Q6s*X1s3OTcuv*WOA\/GoogleAutoML1.png?width=250\" width=\"250\" class=\"align-right\"><\/a>It\u2019s not clear at this point what categories of images will be allowed at launch, but user screens show guidance for general, face, logo, landmarks, and perhaps others.\u00a0 From screen shots shared by Google it appears these models train in the range of about 20 minutes to a few hours.<\/p>\n<p>In the data we were able to find, use appears to be via API.\u00a0 There\u2019s no mention of export code for offline use.\u00a0 Early alpha users include Disney and Urban Outfitters.\u00a0<\/p>\n<p>Anticipating that many new users won\u2019t have labeled data, Google offers access to its own human-labeling services for an additional fee.<\/p>\n<p>Beyond transfer learning, all the majors including Google are pursuing automated ways of automating the optimal tuning of CNNs and RNNs.\u00a0 Handcrafted models are the norm today and are the reason so many often unsuccessful iterations are required.\u00a0<\/p>\n<p><a href=\"http:\/\/api.ning.com\/files\/ds5VTXUtVmst6mxNqM447qwgZwTi9io43Lhekplr*KGKzMvudJ7vIM-ypBAno4AXXeCT457iR2fMNUTXgac9VjmEMgjBdHPX\/GoogleautoMLRNNforoptimization.png\" target=\"_self\"><img decoding=\"async\" src=\"http:\/\/api.ning.com\/files\/ds5VTXUtVmst6mxNqM447qwgZwTi9io43Lhekplr*KGKzMvudJ7vIM-ypBAno4AXXeCT457iR2fMNUTXgac9VjmEMgjBdHPX\/GoogleautoMLRNNforoptimization.png?width=300\" width=\"300\" class=\"align-right\"><\/a>Google calls this next technology Learn2Learn.\u00a0 Currently they are experimenting with RNNs to optimize layers, layer types, nodes, connections, and the other hyperparameters.\u00a0 Since this is basically very high speed random search the compute resources can be extreme.\u00a0<\/p>\n<p>Next on the horizon is the use of evolutionary algorithms to do the same which are much more efficient in terms of time and compute.\u00a0 In a recent presentation, Google researchers showed good results from this approach but they were still taking 3 to 10 days to train just for the optimization.<\/p>\n<\/p>\n<p><span style=\"font-size: 12pt;\"><strong>OneClick.AI<\/strong><\/span><\/p>\n<p>OneClick.AI is an automated machine learning (AML) platform new in the market late in 2017 which includes both traditional algorithms and also deep learning algorithms.<\/p>\n<p>OneClick.AI would be worth a look based on just its AML credentials which include blending, prep, feature engineering, and feature selection, followed by the traditional multi-models in parallel to identify a champion model.<\/p>\n<p>However, what sets OneClick apart is that it includes both image and text DL algos with both transfer learning as well as fully automated hyperparameter tuning for de novo image or text deep learning models.<\/p>\n<p>Unlike Google and Microsoft they are ready to deliver on both image and text.\u00a0 Beyond that, they blend DNNs with traditional algos in ensembles, and use DNNs for forecasting.\u00a0<\/p>\n<p>Forecasting is a little explored area of use for DNNs but it\u2019s been shown to easily outperform other times series forecasters like ARIMA and ARIMAX.<\/p>\n<p>For a platform with this complex offering of tools and techniques it maintains its claim to super easy one-click-data-in-model-out ease which I identify as the minimum requirement for Automated Machine Learning, but which also includes Automated Deep Learning.<\/p>\n<p>The methods used for optimizing its deep learning models are proprietary, but Yuan Shen, Founder and CEO describes it as using AI to train AI, presumably a deep learning approach to optimization.<\/p>\n<\/p>\n<p><span style=\"font-size: 12pt;\"><strong>Which is Better?<\/strong><\/span><\/p>\n<p>It\u2019s much too early to expect much in the way of benchmarking but there is one example to offer, which comes from OneClick.AI.<\/p>\n<p>In a hackathon earlier this year the group tested OneClick against Microsoft\u2019s CustomVision (Google AutoML wasn\u2019t available).\u00a0 Two image classification problems were tested.\u00a0 Tagging photos with:<\/p>\n<p>Horses running or horses drinking water.<\/p>\n<p>\u00a0<a href=\"http:\/\/api.ning.com\/files\/ds5VTXUtVmuNvfYFCfb6gqnTeD6RQgYhNcl72TO1WWShGD6E*KRGfZoJ4gIEW9joQPWUgd7Qo7y2XrzLbc8qLxJunoQO3662\/horsesdrinkingwater.jpg\" target=\"_self\"><img decoding=\"async\" src=\"http:\/\/api.ning.com\/files\/ds5VTXUtVmuNvfYFCfb6gqnTeD6RQgYhNcl72TO1WWShGD6E*KRGfZoJ4gIEW9joQPWUgd7Qo7y2XrzLbc8qLxJunoQO3662\/horsesdrinkingwater.jpg?width=500\" width=\"500\" class=\"align-center\"><\/a><\/p>\n<p>\u00a0Detecting photos with nudity.<\/p>\n<p>\u00a0<a href=\"http:\/\/api.ning.com\/files\/ds5VTXUtVmupJRS6iibKn-vXITHYDSL-XBIPQ0moka6a9w*RGBWbhmuIyUsl6RvcJwiHzqHp0z0xp2WiLkhusfy5a4Naqdk0\/photoswithnudity.jpg\" target=\"_self\"><img decoding=\"async\" src=\"http:\/\/api.ning.com\/files\/ds5VTXUtVmupJRS6iibKn-vXITHYDSL-XBIPQ0moka6a9w*RGBWbhmuIyUsl6RvcJwiHzqHp0z0xp2WiLkhusfy5a4Naqdk0\/photoswithnudity.jpg?width=500\" width=\"500\" class=\"align-center\"><\/a><\/p>\n<p>The horse tagging task was multi-label classification, and the nudity detection task was binary classification. For each task they used 20 images for training, and another 20 for testing.<\/p>\n<ul>\n<li>Horse tagging accuracy: 90% (OneClick.ai) vs. 75% (Microsoft Custom Vision)<\/li>\n<li>Nudity detection accuracy: 95% (OneClick.ai) vs. 50% (Microsoft Custom Vision)<\/li>\n<\/ul>\n<p>This lacks statistical significance and uses only a very small sample in transfer learning.\u00a0 However the results look promising.<\/p>\n<p>This is transfer learning.\u00a0 We\u2019re very interested to see comparisons of the automated model optimization.\u00a0 OneClick\u2019s is ready now.\u00a0 Google should follow shortly.<\/p>\n<p>You may also be asking, where is Amazon in all of this?\u00a0 In our search we couldn\u2019t find any reference to a planned AutoDL offering, but it can\u2019t be far behind.<\/p>\n<p>\u00a0<\/p>\n<p>\u00a0<\/p>\n<p><strong>P<span style=\"font-size: 12pt;\">revious articles on Automated Machine Learning<\/span><\/strong><\/p>\n<p><a href=\"https:\/\/www.datasciencecentral.com\/profiles\/blogs\/next-generation-automated-machine-learning-aml\"><em><u>Next Generation Automated Machine Learning (AML)<\/u><\/em><\/a> (April 2018)<\/p>\n<p><a href=\"https:\/\/www.datasciencecentral.com\/profiles\/blogs\/more-on-fully-automated-machine-learning\"><em><u>More on Fully Automated Machine Learning<\/u><\/em><\/a> \u00a0(August 2017)<\/p>\n<p><a href=\"https:\/\/www.datasciencecentral.com\/profiles\/blogs\/automated-machine-learning-for-professionals\"><em><u>Automated Machine Learning for Professionals<\/u><\/em><\/a> \u00a0(July 2017)<\/p>\n<p><a href=\"https:\/\/www.datasciencecentral.com\/profiles\/blogs\/data-scientists-automated-and-unemployed-by-2025-update\"><em><u>Data Scientists Automated and Unemployed by 2025 &#8211; Update!<\/u><\/em><\/a> \u00a0(July 2017)<\/p>\n<p><a href=\"https:\/\/www.datasciencecentral.com\/profiles\/blogs\/data-scientists-automated-and-unemployed-by-2025\"><em><u>Data Scientists Automated and Unemployed by 2025!<\/u><\/em><\/a>\u00a0 (April 2016)<\/p>\n<p>\u00a0<\/p>\n<p>\u00a0<\/p>\n<p>About the author:\u00a0 Bill Vorhies is Editorial Director for Data Science Central and has practiced as a data scientist since 2001.\u00a0 He can be reached at:<\/p>\n<p><a href=\"mailto:Bill@DataScienceCentral.com\">Bill@DataScienceCentral.com<\/a><\/p>\n<\/div>\n<p><a href=\"https:\/\/www.datasciencecentral.com\/xn\/detail\/6448529:BlogPost:711828\">Go to Source<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Author: William Vorhies Summary:\u00a0 There are several things holding back our use of deep learning methods and chief among them is that they are complicated [&hellip;] <span class=\"read-more-link\"><a class=\"read-more\" href=\"https:\/\/www.aiproblog.com\/index.php\/2018\/10\/09\/automated-deep-learning-so-simple-anyone-can-do-it\/\">Read More<\/a><\/span><\/p>\n","protected":false},"author":1,"featured_media":467,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_bbp_topic_count":0,"_bbp_reply_count":0,"_bbp_total_topic_count":0,"_bbp_total_reply_count":0,"_bbp_voice_count":0,"_bbp_anonymous_reply_count":0,"_bbp_topic_count_hidden":0,"_bbp_reply_count_hidden":0,"_bbp_forum_subforum_count":0,"footnotes":""},"categories":[26],"tags":[],"_links":{"self":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/1136"}],"collection":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/comments?post=1136"}],"version-history":[{"count":0,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/1136\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media\/457"}],"wp:attachment":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media?parent=1136"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/categories?post=1136"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/tags?post=1136"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}