{"id":2458,"date":"2019-08-13T06:30:30","date_gmt":"2019-08-13T06:30:30","guid":{"rendered":"https:\/\/www.aiproblog.com\/index.php\/2019\/08\/13\/doctors-are-from-venus-data-scientists-from-mars-or-why-ai-ml-is-moving-so-slowly-in-healthcare\/"},"modified":"2019-08-13T06:30:30","modified_gmt":"2019-08-13T06:30:30","slug":"doctors-are-from-venus-data-scientists-from-mars-or-why-ai-ml-is-moving-so-slowly-in-healthcare","status":"publish","type":"post","link":"https:\/\/www.aiproblog.com\/index.php\/2019\/08\/13\/doctors-are-from-venus-data-scientists-from-mars-or-why-ai-ml-is-moving-so-slowly-in-healthcare\/","title":{"rendered":"Doctors are from Venus, Data Scientists from Mars \u2013 or Why AI\/ML is Moving so Slowly in Healthcare"},"content":{"rendered":"<p>Author: William Vorhies<\/p>\n<div>\n<p><strong><em>Summary:<\/em><\/strong> <em>The world of healthcare may look like the most fertile field for AI\/ML apps but in practice it\u2019s fraught with barriers.\u00a0 These range from cultural differences, to the failure of developers to really understand the environment they are trying to enhance, to regulatory and logical Catch 22s that work against adoption.\u00a0 Part 3 of 3.<\/em><\/p>\n<p>\u00a0<\/p>\n<p><a href=\"https:\/\/storage.ning.com\/topology\/rest\/1.0\/file\/get\/831158652?profile=original\" target=\"_blank\" rel=\"noopener noreferrer\"><img decoding=\"async\" src=\"https:\/\/storage.ning.com\/topology\/rest\/1.0\/file\/get\/831158652?profile=RESIZE_710x\" width=\"300\" class=\"align-right\"><\/a>According to data compiled by research firm Startup Health funding for digital healthcare totaled $14.6 billion in 2018.\u00a0 The world of healthcare may look like the most fertile field for AI\/ML apps but in practice it\u2019s fraught with barriers.\u00a0 These range from cultural differences, to the failure of developers to really understand the environment they are trying to enhance, to regulatory and logical Catch 22s that work against adoption.\u00a0<\/p>\n<p>This is part 3 of our three part series on AI\/ML in healthcare.\u00a0 The content is the result of my attending the <a href=\"http:\/\/ai-med.io\/\"><em><u>AIMed Conference<\/u><\/em><\/a> last December which is unique because its attendees are 80% clinicians and hospital CIO\/Administrators and not data scientists.<\/p>\n<p>The really unique value of this conference is seeing the AI\/ML landscape through the eyes of users, and not through the overly optimistic eyes of data scientists and VCs.\u00a0<\/p>\n<p>In <a href=\"https:\/\/www.datasciencecentral.com\/profiles\/blogs\/seeing-the-ai-ml-future-in-healthcare-through-the-eyes-of-physici\"><em><u>part 1<\/u><\/em><\/a> we talked about the extremely low adoption rate of AI\/ML in healthcare, on the order of only 1% of hospitals.\u00a0<\/p>\n<p>In <a href=\"https:\/\/www.datasciencecentral.com\/profiles\/blogs\/the-ai-ml-opportunity-landscape-in-healthcare-do-it-right-or-it-w\"><em><u>part 2<\/u><\/em><\/a> we tried to set those opportunities into an orderly landscape organized around the physician and their patients.\u00a0<\/p>\n<p>In part 3 we\u2019ll describe what those physicians and administrators told us that\u2019s holding back adoption.<\/p>\n<p>Here are some of the many reasons stated by clinicians attending the AIMed Conference.<\/p>\n<p>\u00a0<\/p>\n<p><span style=\"font-size: 12pt;\"><strong>Too Many False Positives<\/strong><\/span><\/p>\n<p>Data scientists don\u2019t need to think twice about the fact that all of our techniques are probabilistic and contain both false positive and false negative errors.<\/p>\n<p>In healthcare however, false negatives, that is failing to detect a disease state is the ultimate failure to be avoided at all cost.\u00a0 As a result, applications designed, for example, to automatically detect cancer or other diseases in medical images are tuned to minimize these type 2 errors.<\/p>\n<p>This necessarily increases false positives that can only be reduced by increasing overall model accuracy.\u00a0 And that can only happen where a large amount of training data is available.\u00a0 More on that later.<\/p>\n<p>Radiologists and pathologists complain that false positives slow them down too much as they are forced to examine all the portions of the image flagged by the model.\u00a0 And in fact spend even more time on the false positive indications, not wanting to miss something important.\u00a0<\/p>\n<p><a href=\"https:\/\/storage.ning.com\/topology\/rest\/1.0\/file\/get\/831176634?profile=original\" target=\"_blank\" rel=\"noopener noreferrer\"><img decoding=\"async\" src=\"https:\/\/storage.ning.com\/topology\/rest\/1.0\/file\/get\/831176634?profile=RESIZE_710x\" width=\"250\" class=\"align-right\"><\/a>The radiologists who spoke at the conference said the impact was so severe that not only was there no time savings but the use of these systems actually took more time to evaluate.\u00a0 The model-based system may be faster, but the total time to evaluate was longer.<\/p>\n<p>One solution suggested, aside from diminishing error, was for the model to describe specifically what caused a particular area to be called out as abnormal, for example by identifying a particular type of cell abnormality that the trained radiologist or pathologist would recognize.\u00a0 Think of this as the clinician\u2019s version of a plea for transparency.<\/p>\n<p>Similarly with the many IoT type applications being promoted to monitor in-patients for critical events, clinicians, nursing, and other professional staff reported \u2018<strong>alarm fatigue<\/strong>\u2019 from too many false positives, reducing the likelihood that they would respond with urgency.<\/p>\n<p>\u00a0<\/p>\n<p><span style=\"font-size: 12pt;\"><strong>Turn Down the Hype<\/strong><\/span><\/p>\n<p>While we\u2019re on the topic of automated image evaluation, radiologists and pathologists would like the press to turn down the hype on these \u2018breakthroughs\u2019 often described as new levels of accuracy in the detection of this or that cancer.\u00a0<\/p>\n<p>They remind us that the job of radiologists and pathologists is not to tell the treating doctor that they have discovered a cancer in the image, but rather to say that a specific area looks suspicious and requires the doctor to examine it more closely.<\/p>\n<p>\u00a0<\/p>\n<p><span style=\"font-size: 12pt;\"><strong>Don\u2019t Disrupt my Workflow Bro<\/strong><\/span><\/p>\n<p><a href=\"https:\/\/storage.ning.com\/topology\/rest\/1.0\/file\/get\/831207318?profile=original\" target=\"_blank\" rel=\"noopener noreferrer\"><img decoding=\"async\" src=\"https:\/\/storage.ning.com\/topology\/rest\/1.0\/file\/get\/831207318?profile=RESIZE_710x\" width=\"350\" class=\"align-right\"><\/a>Unlike the human workers in the broader world of commerce, hospitals are staffed by a very high percentage of very highly educated workers (clinicians) who seem to be always scheduled to the very edge of efficiency or exhaustion, depending on your point of view.<\/p>\n<p>It\u2019s in the nature of running a hospital not to have too few or too many of any particular clinical specialty for reasons of cost.\u00a0 That means that healthcare professionals are quite possibly the most overworked or at least critically scheduled category of workers anywhere.\u00a0 There is seldom a moment when they are under-utilized.\u00a0<\/p>\n<p>Where in the world of general commerce for example would you have your specialist sleep at the office so you could wake him in the middle of the night to evaluate a problem.<\/p>\n<p>The result is the adoption of natural workflow patterns that allow attending physicians to see as many patients as possible (without causing harm) or for radiologists and pathologists to examine as many images or slides as possible in the shortest amount of time.\u00a0 For example, the average radiologist is said to serve 200 patients per day, evaluating 3,000 medical images at least 90% of which will be normal.<\/p>\n<p>These folks have all learned and developed workflow techniques that maximize their effectiveness and efficiency.\u00a0 It\u2019s the heart of a unique culture that is the opposite of the equally unique culture of the healthcare data science startup to disrupt the status quo with their innovative breakthrough de jour.<\/p>\n<p>For example, in automated image classification, some solutions have attempted to move radiologists off of their film-based media and pathologists off of their microscope-slide based tools by offering large hi-def holographic screens instead.\u00a0 This seemingly harmless improvement was quoted as the source of work slowdown that caused these solutions to be rejected.<\/p>\n<p>This is at the heart of the Mars\/Venus analogy in the title.\u00a0 What we heard repeatedly is that new innovations need to seamlessly integrate with existing workflows and practices.\u00a0 Evidently this obvious UX element is missing in many proposed solutions.\u00a0 For adoption to take place, data scientists need to understand the culture, and particularly the importance of integrating with current workflow and practices.<\/p>\n<p>\u00a0<\/p>\n<p><span style=\"font-size: 12pt;\"><strong>FOSS versus FOMO<\/strong><\/span><\/p>\n<p>By the way, FOSS (fear of small startups) has more weight than FOMO (fear of missing out).\u00a0 Even where AI\/ML solutions like automated image classification were shown to be promising, hospital administrators showed the same reticence to contract with new, small startups that any competent commercial enterprise would.<\/p>\n<p>When you are investing in an embedded AI\/ML solution you are betting that the vendor will be around to continue to support and upgrade that solution.\u00a0 Small startups in any industry have the obvious risk of not surviving.<\/p>\n<p>Administrators and clinicians asked that embedded AI\/ML solution providers, particularly pointing at the automated imaging solutions, should partner with the \u2018hard metal\u2019 machine providers to prove or at least insure their staying power.<\/p>\n<p>\u00a0<\/p>\n<p><span style=\"font-size: 12pt;\"><strong>The Electronic Healthcare Record (EHR) \u2013 A Deal with the Devil<\/strong><\/span><\/p>\n<p>A survey from the AMA this year continues to show that the EHR and related clinical systems are the chief reason for physician burnout.\u00a0 It\u2019s widely quoted that the introduction of these systems has burdened the physician with two hours of administration for every hour of patient-facing time.<\/p>\n<p>If you\u2019ve been to your doctor recently you know that the keyboard is now an almost physical barrier between your physician and you.\u00a0 It\u2019s reported that this is the first generation of physicians NOT to recommend that their children enter medicine.<\/p>\n<p>And yet to get to the benefits of AI\/ML in healthcare requires the data that starts here in the EHR.<\/p>\n<p>There are many structural and procedural problems with health data but key among them is extracting that data from these electronic health records.<\/p>\n<p>China has an initiative underway employing 50,000 medical students to extract and transcribe this data into data bases.\u00a0 But the feeling among US clinicians is that medical students aren\u2019t qualified to extract this data accurately.\u00a0 That it needs to happen at the time of data capture.<\/p>\n<p>This practically screams out NLP to any data scientist and some applications are making inroads.\u00a0 However, some still haven\u2019t learned lesson 2 above about integrating into existing workflows.<\/p>\n<p>For example, for both data capture and for liability reasons, hospitals would like to have step-by-step documentation of what happens during a surgery.\u00a0 One proposed solution was to put an elaborate headset on the surgeon that would both record video and allow the surgeon to dictate his actions.\u00a0 Guess how well that was accepted.<\/p>\n<p>A major challenge to NLP solutions and indeed all types of data capture in healthcare is interoperability among different data sets.\u00a0 The consistency and standardization isn\u2019t there today restricting most data sets to relatively small size and making the blending of data sets chancy at best.<\/p>\n<p>This is a huge pain point and barrier where NLP promises improvement and where movement is underway toward the standardization necessary.\u00a0 It\u2019s not there yet.<\/p>\n<p>\u00a0<\/p>\n<p><span style=\"font-size: 12pt;\"><strong>Data is Too Thin and Won\u2019t Generalize<\/strong><\/span><\/p>\n<p><a href=\"https:\/\/storage.ning.com\/topology\/rest\/1.0\/file\/get\/831218654?profile=original\" target=\"_blank\" rel=\"noopener noreferrer\"><img decoding=\"async\" src=\"https:\/\/storage.ning.com\/topology\/rest\/1.0\/file\/get\/831218654?profile=RESIZE_710x\" width=\"300\" class=\"align-right\"><\/a>The problems with the healthcare data necessary to train AI\/ML solutions doesn\u2019t end with extraction.\u00a0 The first major problem is that the data is simply too thin and won\u2019t generalize.<\/p>\n<p>There are a few large data bases in the 100,000 record range but the effort to rollup patient data into data science worthy DBs is early in the process.\u00a0 Some of the obvious problems:<\/p>\n<ol>\n<li>A solution from one country or even one hospital won\u2019t necessarily generalize to other populations, inside or outside of the US.<\/li>\n<li>Both privacy regulations and the feeling on the part of many hospitals that they should be compensated for their data are dramatically limiting sharing. For example, although medical imaging accounts for 90% of all healthcare data, 97% goes unanalyzed or unused per Keith Bigelow, SVP for GE Healthcare.<\/li>\n<li>In clinical informatics, the first step in the AI\/ML augmented physician, creating models to predict better outcomes or prevent worse outcomes are being attempted. But data is so thin at the hospital level that the models suffer from what data scientists understand as \u2018leakage from the future\u2019.\u00a0 That is, when the first promising results are found and practices are changed, the data from the new group impacted by the model is mixed back with original data eliminating a proper control group.\u00a0 Still, having immediate benefit is seen as preferable.<\/li>\n<\/ol>\n<p>\u00a0<\/p>\n<p><span style=\"font-size: 12pt;\"><strong>Continuous Learning is Broken<\/strong><\/span><\/p>\n<p>What we all want is to continuously improve our models.\u00a0 Thin data resulting in blending new and old data and suffering leakage from the future is only part of this.<\/p>\n<p>The siloing or hoarding of data at the hospital level is reported to be an equal problem.\u00a0 When a hospital accepts an embedded AI\/ML solution from a new vendor, there are several factors that may prevent results data from flowing back for improvement.\u00a0<\/p>\n<p>One issue is interoperability even at the machine level.\u00a0 Several examples have been shown in imaging where data from similar machines with different manufacture or even different settings are not comparable.\u00a0<\/p>\n<p>Second, some hospitals take the position that they are financial partners in this arrangement and need to be compensated for the returning data.<\/p>\n<p>An even more formidable barrier is raised by the FDA in approving imaging based solutions.\u00a0 On the one hand, the FDA has taken a very permissive approach to approving AI\/ML image classifying solutions based on training with as few as 100 to 300 images.\u00a0<\/p>\n<p>However, those approvals are then frozen and require reapplication before an improved solution can be released.\u00a0 Not to mention that using such a small number of images for training implies transfer learning and inaccuracies that may arise from using a base model that actually does not transfer weights well in differing circumstances.<\/p>\n<p>\u00a0<\/p>\n<p><span style=\"font-size: 12pt;\"><strong>Not So Fast With Those Rollouts<\/strong><\/span><\/p>\n<p>This problem lies squarely at the feet of those data science healthcare vendors still in the \u2018move fast and break things\u2019 mindset.\u00a0 The best example is the use of chatbots which are a natural for patient scheduling, intake, and even determining when or whether a patient should see a doctor.<\/p>\n<p>In a widely touted rollout just a few months ago, a chatbot called Babylon was rolled out in the UK to provide diagnostic advice on common ailments without human interaction.\u00a0 The app would then also vet access to the single-payer UK system presumably reducing the cost of the initial interaction and offering accurate and timely referrals to the correct physicians and hospitals.<\/p>\n<p>However, as reported in a <a href=\"https:\/\/www.forbes.com\/sites\/parmyolson\/2018\/12\/17\/this-health-startup-won-big-government-dealsbut-inside-doctors-flagged-problems\/?utm_campaign=the_download.unpaid.engagement&#038;utm_source=hs_email&#038;utm_medium=email&#038;utm_content=68457957&#038;_hsenc=p2ANqtz-8bBqTQS18aIfduCD6IWNpl-9-z6PEtaqbNwvZZo304a-kHgnlWWDFatHLfEDXbMhP2Rp5uqtPz02E6WvJj0PIxqLu-tA&#038;_hsmi=68457957#1ae12195eabb\"><em><u>recent Forbes article<\/u><\/em><\/a>, a group of auditing physicians \u201cfound that around 10% to 15% of the chatbot\u2019s 100 most frequently suggested outcomes, such as a chest infection, either missed warning signs of a more serious condition like cancer or sepsis or were just flat-out wrong\u201d.<\/p>\n<p>The problem was simply shortcuts taken during training and too great an emphasis on rolling out fast before being audited.<\/p>\n<p>It\u2019s likely that most non-data scientists <a href=\"https:\/\/www.datasciencecentral.com\/profiles\/blogs\/beginners-guide-to-chatbots\"><em><u>think that chatbots are actually smarter than they are<\/u><\/em><\/a>.\u00a0 The AI in chatbot refers to its NLP ability to understand free form text and voice input and create similar output.<\/p>\n<p>What most folks don\u2019t realize is that the internal logic of which answers to provide are still hand coded decision trees in 95% of chatbots, not the result of some exotic AI\/ML related search or automated intelligence (which might not be better anyway).<\/p>\n<p>So buyers beware and be sure to satisfy yourself about the accuracy of any chatbot or similar AI\/ML solutions before you put them in production.\u00a0 That\u2019s one of the first lessons we learn in commercial rollouts.<\/p>\n<p>\u00a0<\/p>\n<p><span style=\"font-size: 12pt;\"><strong>Finally<\/strong><\/span><\/p>\n<p>One of the conference speakers, Ted Shortliffe who is a physician, data scientist, and widely credited author observed that perhaps the reason for slow adoption in healthcare has to do with the fact that it is a \u2018real physical business\u2019.\u00a0<\/p>\n<p>He contrasts this with where AI\/ML had its earliest successes in the e-world of finance, ecommerce, and entertainment, none of which is hobbled by an existing physical operational environment.\u00a0<\/p>\n<p>Perhaps he\u2019s right that the brick and mortar world is a much tougher nut to crack and requires a slower, more deliberate approach.\u00a0 Especially when complicated with the specialized processes and skills needed in healthcare.<\/p>\n<p>For data scientists hoping to capitalize in this market, there are few important lessons.\u00a0<\/p>\n<ul>\n<li>Slow down a little and make sure you understand how your disruptive application can actually be integrated into this world of specialized workflows.<\/li>\n<li>Make sure you understand both the current restrictions on data size and accuracy, and how long it may take before that gets better.<\/li>\n<li>Don\u2019t rush the rollout. People\u2019s health is at stake.<\/li>\n<\/ul>\n<p>And if your VC is pushing you a little too hard, remember that the early bird may get the worm, but the second mouse gets the cheese.<\/p>\n<p>\u00a0<\/p>\n<p><strong>Other Articles in this Series<\/strong><\/p>\n<p><a href=\"https:\/\/www.datasciencecentral.com\/profiles\/blogs\/seeing-the-ai-ml-future-in-healthcare-through-the-eyes-of-physici\"><em><u>Seeing the AI\/ML Future in Healthcare Through the Eyes of Physicians. It Doesn\u2019t Look Quite the Way Data Scientists See It<\/u><\/em><\/a><\/p>\n<p><a href=\"https:\/\/www.datasciencecentral.com\/profiles\/blogs\/the-ai-ml-opportunity-landscape-in-healthcare-do-it-right-or-it-w\"><em><u>The AI\/ML Opportunity Landscape in Healthcare. Do It Right or It Will be More of a Mine Field<\/u><\/em><\/a><\/p>\n<p>\u00a0<\/p>\n<p><a href=\"https:\/\/www.datasciencecentral.com\/profiles\/blog\/list?user=0h5qapp2gbuf8\"><em><u>Other articles by Bill Vorhies.<\/u><\/em><\/a><\/p>\n<p><em><u>\u00a0<\/u><\/em><\/p>\n<p>About the author:\u00a0 Bill is Editorial Director for Data Science Central.\u00a0 Bill is also President &#038; Chief Data Scientist at Data-Magnum and has practiced as a data scientist since 2001.\u00a0\u00a0\u00a0 He can be reached at:<\/p>\n<p><a href=\"mailto:Bill@DataScienceCentral.com\">Bill@DataScienceCentral.com<\/a> <span>or<\/span> <a href=\"mailto:Bill@Data-Magnum.com\">Bill@Data-Magnum.com<\/a><\/p>\n<\/div>\n<p><a href=\"https:\/\/www.datasciencecentral.com\/xn\/detail\/6448529:BlogPost:794759\">Go to Source<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Author: William Vorhies Summary: The world of healthcare may look like the most fertile field for AI\/ML apps but in practice it\u2019s fraught with barriers.\u00a0 [&hellip;] <span class=\"read-more-link\"><a class=\"read-more\" href=\"https:\/\/www.aiproblog.com\/index.php\/2019\/08\/13\/doctors-are-from-venus-data-scientists-from-mars-or-why-ai-ml-is-moving-so-slowly-in-healthcare\/\">Read More<\/a><\/span><\/p>\n","protected":false},"author":1,"featured_media":463,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_bbp_topic_count":0,"_bbp_reply_count":0,"_bbp_total_topic_count":0,"_bbp_total_reply_count":0,"_bbp_voice_count":0,"_bbp_anonymous_reply_count":0,"_bbp_topic_count_hidden":0,"_bbp_reply_count_hidden":0,"_bbp_forum_subforum_count":0,"footnotes":""},"categories":[26],"tags":[],"_links":{"self":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/2458"}],"collection":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/comments?post=2458"}],"version-history":[{"count":0,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/posts\/2458\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media\/461"}],"wp:attachment":[{"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/media?parent=2458"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/categories?post=2458"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.aiproblog.com\/index.php\/wp-json\/wp\/v2\/tags?post=2458"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}