{"id":58,"date":"2020-03-18T12:38:27","date_gmt":"2020-03-18T12:38:27","guid":{"rendered":"https:\/\/sites.mitre.org\/aifails\/?page_id=58"},"modified":"2020-07-13T08:42:50","modified_gmt":"2020-07-13T12:42:50","slug":"after-developing-the-ai","status":"publish","type":"page","link":"https:\/\/sites.mitre.org\/aifails\/after-developing-the-ai\/","title":{"rendered":"We&#8217;re Not Done Yet: After Developing the AI"},"content":{"rendered":"\n<p>[et_pb_section fb_built=&#8221;1&#8243; specialty=&#8221;on&#8221; _builder_version=&#8221;4.2.2&#8243;][et_pb_column type=&#8221;1_4&#8243; _builder_version=&#8221;3.25&#8243; custom_padding=&#8221;|||&#8221; custom_padding__hover=&#8221;|||&#8221;][et_pb_sidebar area=&#8221;et_pb_widget_area_19&#8243; admin_label=&#8221;After Developing the AI&#8221; _builder_version=&#8221;4.4.1&#8243; min_height=&#8221;200px&#8221; custom_margin=&#8221;||0px||false|false&#8221; custom_padding=&#8221;||0px||false|false&#8221;][\/et_pb_sidebar][\/et_pb_column][et_pb_column type=&#8221;3_4&#8243; specialty_columns=&#8221;3&#8243; _builder_version=&#8221;3.25&#8243; custom_padding=&#8221;|||&#8221; custom_padding__hover=&#8221;|||&#8221;][et_pb_row_inner _builder_version=&#8221;4.0.3&#8243; custom_padding=&#8221;||0px||false|false&#8221;][et_pb_column_inner saved_specialty_column_type=&#8221;3_4&#8243; _builder_version=&#8221;4.0.3&#8243;][et_pb_text admin_label=&#8221;We&#8217;re Not Done Yet: After Developing the AI&#8221; _builder_version=&#8221;4.4.8&#8243; custom_margin=&#8221;||0px||false|false&#8221; custom_padding=&#8221;5px||0px||false|false&#8221;]<\/p>\n<h1><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-3832 size-full alignleft\" src=\"https:\/\/sites.mitre.org\/aifails\/wp-content\/uploads\/sites\/15\/2020\/02\/icon_developing-002.png\" alt=\"\" width=\"135\" height=\"120\" \/>We&#8217;re Not Done Yet: After Developing the AI<\/h1>\n<p>Developing AI is a dynamic, multifaceted process. Even if an AI performs optimally from a technical standpoint, other constraining factors could limit its overall performance and acceptance. Developing an AI to be safe and dependable means stakeholders must learn more about how the AI functions as the risks from its use increase. This section details factors that make that understanding challenging to achieve, and describes how proper documentation, explanations of intent, and user education can improve outcomes.<\/p>\n<p>&nbsp;<\/p>\n<p>[\/et_pb_text][et_pb_text admin_label=&#8221;Fails&#8221; _builder_version=&#8221;4.4.8&#8243; custom_margin=&#8221;||0px||false|false&#8221; custom_padding=&#8221;5px||0px||false|false&#8221;]<\/p>\n<div id=\"developing\">\n\u00a0\n<\/div>\n<h5>Explore the Three Fails in This Category:<\/h5>\n<p>[\/et_pb_text][et_pb_tabs active_tab_background_color=&#8221;#a0ddf3&#8243; inactive_tab_background_color=&#8221;#e5e7e8&#8243; active_tab_text_color=&#8221;#000000&#8243; admin_label=&#8221;Fails and Lessons Learned&#8221; module_class=&#8221;icon-tabs&#8221; _builder_version=&#8221;4.4.8&#8243; tab_text_color=&#8221;#000000&#8243; body_text_color=&#8221;#000000&#8243; tab_font_size=&#8221;15px&#8221; tab_line_height=&#8221;1.3em&#8221; custom_margin=&#8221;|||0px|false|false&#8221; custom_padding=&#8221;|||0px|false|false&#8221;][et_pb_tab title=&#8221;Testing in the Wild&#8221; _builder_version=&#8221;4.4.8&#8243;]<\/p>\n<div style=\"padding: 0px 0px 15px 0px;margin: 0px 0px 0px 0px\">\n<h3>Testing in the Wild<\/h3>\n<p>Test and evaluation (T&amp;E) teams work with algorithm developers to outline criteria for quality control, and of course they can\u2019t anticipate all algorithmic outcomes. But the consequences (and even blame) for the unexpected results are sometimes transferred onto groups who are unaware of these limitations or have not consented to being test subjects.<\/p>\n<p>&nbsp;<\/p>\n<\/div>\n<div class=\"example-callout\">\n<p><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-4030 size-full alignleft z-index40\" src=\"https:\/\/sites.mitre.org\/aifails\/wp-content\/uploads\/sites\/15\/2020\/03\/icon_example_robot-002.png\" alt=\"\" width=\"100\" height=\"83\" \/><\/p>\n<h3>Examples<\/h3>\n<p>Boeing initially blamed foreign pilots for the 737 MAX crashes, even though a sensor malfunction, faulty software, lack of pilot training, making a safety feature an optional purchase, and not mentioning the software in the pilot manual were all contributory causes.<a class=\"reference\" href=\".\/references\/#11.1\" target=\"_blank\" rel=\"noopener noreferrer\">1<\/a><\/p>\n<p>In 2014, UK immigration rules required some foreigners to pass an English proficiency test. A voice recognition system was used as part of the exam to detect fraud (e.g., if an applicant took the test multiple times under different names, or if a native speaker took the oral test posing as the applicant). But because the government did not understand how high the algorithm\u2019s error rate was, and each flagged recording was checked by undertrained employees, the UK cancelled thousands of visas and deported people in error.<a class=\"reference\" href=\".\/references\/#11.2\" target=\"_blank\" rel=\"noopener noreferrer\">2,<\/a><a class=\"reference\" href=\".\/references\/#11.3\" target=\"_blank\" rel=\"noopener noreferrer\">3<\/a> Thus, applicants who had followed the rules suffered the consequences of the shortcomings in the algorithm.<\/p>\n<p><a href=\"#_ednref1\" name=\"_edn1\" id=\"_edn1\"><\/a><\/p>\n<\/div>\n<h3>Why is this a fail?<\/h3>\n<p>T&amp;E of AI algorithms is hard. Even for AI models that aren\u2019t entirely black boxes we have only limited T&amp;E tools<a class=\"reference\" href=\".\/references\/#11.4\" target=\"_blank\" rel=\"noopener noreferrer\">4,<\/a><a class=\"reference\" href=\".\/references\/#11.5\" target=\"_blank\" rel=\"noopener noreferrer\">5<\/a> (though resources are emerging<a class=\"reference\" href=\".\/references\/#11.6\" target=\"_blank\" rel=\"noopener noreferrer\">6,<\/a><a class=\"reference\" href=\".\/references\/#11.7\" target=\"_blank\" rel=\"noopener noreferrer\">7,<\/a><a class=\"reference\" href=\".\/references\/#11.8\" target=\"_blank\" rel=\"noopener noreferrer\">8<\/a>). Difficulties for T&amp;E result from:<\/p>\n<p><a href=\"#_ednref1\" name=\"_edn1\"><\/a><em>Uncertain outcomes:<\/em> Many AI models are complex, not fully explainable, and potentially non-linear (meaning they behave in unexpected ways in response to unexpected inputs), and we don\u2019t have great tools to help us understand their decisions and limitations.<a class=\"reference\" href=\".\/references\/#11.9\" target=\"_blank\" rel=\"noopener noreferrer\">9,<\/a><a class=\"reference\" href=\".\/references\/#11.a\" target=\"_blank\" rel=\"noopener noreferrer\">10,<\/a><a class=\"reference\" href=\".\/references\/#11.b\" target=\"_blank\" rel=\"noopener noreferrer\">11<\/a><\/p>\n<p><em>Model drift: <\/em>Due to changes in data, the environment, or people\u2019s behavior an AI\u2019s performance will drift, or become outdated, over time.<a class=\"reference\" href=\".\/references\/#11.c\" target=\"_blank\" rel=\"noopener noreferrer\">12,<\/a><a class=\"reference\" href=\".\/references\/#11.d\" target=\"_blank\" rel=\"noopener noreferrer\">13<\/a><\/p>\n<p><em>Unanticipated use: <\/em>Because AI interacts with people who probably do not share our skills or understanding of the system, and who may not share our goals, the AI will be used in unanticipated ways.<\/p>\n<p><em>Pressures to move quickly:<\/em> There is a tension between resolving to develop and deploy automated products quickly and taking time to test, understand, and address the limitations of those products.<a class=\"reference\" href=\".\/references\/#11.e\" target=\"_blank\" rel=\"noopener noreferrer\">14<\/a><\/p>\n<p>Because all these difficulties, deployers and consumers of AI models often don\u2019t know the range or severity of consequences of the AI\u2019s application.<a class=\"reference\" href=\".\/references\/#11.f\" target=\"_blank\" rel=\"noopener noreferrer\">15<\/a><\/p>\n<p>Jonathan Zittrain, Harvard Law School professor, describes how the issues that emerge from an unpredictable system will become problematic as the number of systems increases. He introduces the concept of \u201cintellectual debt,\u201d which applies to many fields, not only AI. For example, in medicine some drugs are approved for wide use even when \u201cno one knows exactly how they work,\u201d<a class=\"reference\" href=\".\/references\/#11.g\" target=\"_blank\" rel=\"noopener noreferrer\">16<\/a> but they may still have value. If the unknowns were limited to only a single AI (or drug), then causes and effects might be isolated and mitigated. But as the number of AIs and their interactions with humans grows, performing the number of tests required to uncover potential consequences becomes logistically impossible.<\/p>\n<p>&nbsp;<\/p>\n<h3>What happens when things fail?<\/h3>\n<p>Users are held responsible for bad AI outcomes even if those outcomes aren\u2019t entirely (or at all) their fault. A lack of laws defining accountability and responsibility for AI means that it is too easy to blame the AI victim when something goes wrong. The default assumption in semi-autonomous vehicle crashes, as in the Boeing 737 MAX tragedies, has been that drivers are solely at fault.<a class=\"reference\" href=\".\/references\/#11.h\" target=\"_blank\" rel=\"noopener noreferrer\">17,<\/a><a class=\"reference\" href=\".\/references\/#11.i\" target=\"_blank\" rel=\"noopener noreferrer\">18,<\/a><a class=\"reference\" href=\".\/references\/#11.j\" target=\"_blank\" rel=\"noopener noreferrer\">19,<\/a><a class=\"reference\" href=\".\/references\/#11.k\" target=\"_blank\" rel=\"noopener noreferrer\">20,<\/a><a class=\"reference\" href=\".\/references\/#11.l\" target=\"_blank\" rel=\"noopener noreferrer\">21<\/a> Similarly, reports on the 737 crashes showed that \u201call the risk [was put] on the pilot, who would be expected to know what to do within seconds if a system he didn\u2019t know existed\u2026 forced the plane downward.\u201d<a class=\"reference\" href=\".\/references\/#11.m\" target=\"_blank\" rel=\"noopener noreferrer\">22<\/a> The early days of automated flying demonstrated that educating pilots about the automation capabilities and how to act as a member of a human-machine team reduced the number of crashes significantly.<a class=\"reference\" href=\".\/references\/#11.n\" target=\"_blank\" rel=\"noopener noreferrer\">23,<\/a><a class=\"reference\" href=\".\/references\/#11.o\" target=\"_blank\" rel=\"noopener noreferrer\">24,<\/a><a class=\"reference\" href=\".\/references\/#11.p\" target=\"_blank\" rel=\"noopener noreferrer\">25<\/a><\/p>\n<p>As a separate concern, the individuals or communities subject to an AI can become unwilling or unknowing test subjects. Pedestrians can unknowingly be injured by still-learning, semi-autonomous vehicles;<a class=\"reference\" href=\".\/references\/#11.q\" target=\"_blank\" rel=\"noopener noreferrer\">26<\/a> oncology patients can be diagnosed by an experimental IBM Watson (Watson is in a trial phase and not yet approved for clinical use);<a class=\"reference\" href=\".\/references\/#11.r\" target=\"_blank\" rel=\"noopener noreferrer\">27<\/a> Pearson can offer different messaging to different students as an experiment in gauging student engagement.<a class=\"reference\" href=\".\/references\/#11.s\" target=\"_blank\" rel=\"noopener noreferrer\">28<\/a> As the AI Now Institute at New York University (a research institute dedicated to understanding the social implications of AI technologies) puts it, \u201cthis is a repeated pattern when market dominance and profits are valued over safety, transparency, and assurance.\u201d<a class=\"reference\" href=\".\/references\/#11.t\" target=\"_blank\" rel=\"noopener noreferrer\">29<\/a><a href=\"#_ednref11\" name=\"_edn11\" id=\"_edn11\"><\/a><\/p>\n<blockquote>\n<p>The early days of automated flying demonstrated that educating pilots about the automation capabilities and how to act as a member of a human-machine team reduced the number of crashes significantly<\/p>\n<\/blockquote>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<table style=\"border: 5px solid #336699\">\n<tbody>\n <tr>\n  <th class=\"lessonshdr\" colspan=\"4\"><\/th>\n <\/tr>\n <tr>\n  <th class=\"lessons2hdr\" colspan=\"4\"><\/th>\n <\/tr>\n <tr>\n  <th class=\"categoryhdr1\"><\/th>\n  <th class=\"categoryhdr2\"><\/th>\n  <th class=\"categoryhdr3\"><\/th>\n  <th class=\"categoryhdr4\"><\/th>\n <\/tr>\n <tr>\n  <td class=\"inactive\"><a class=\"popmake-3347\" href=\"#\">Hold AI to a Higher Standard<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3405\" href=\"#\">Involve the Communities Affected by the AI <\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3421\" href=\"#\">Make Our Assumptions Explicit<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-2906\" href=\"#\">Monitor the AI&#8217;s Impact and Establish Layers of Accountability<\/a><\/td>\n <\/tr>\n <tr>\n  <td class=\"inactive\"><a class=\"popmake-3096\" href=\"#\">It&#8217;s OK to Say No to Automation<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3409\" href=\"#\">Plan to Fail<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3450\" href=\"#\">Try Human-AI Couples Counseling<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3552\" href=\"#\">Envision Safeguards for AI Advocates<\/a><\/td>\n <\/tr>\n <tr>\n  <td class=\"active\"><a class=\"popmake-3393\" href=\"#\">AI Challenges are Multidisciplinary, so They Require a Multidisciplinary Team<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3414\" href=\"#\">Ask for Help: Hire a Villain<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-4413\" href=\"#\">Offer the User Choices<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3146\" href=\"#\">Require Objective, Third-party Verification and Validation\u00a0<\/a><\/td>\n <\/tr>\n <tr style=\"border-bottom: 5px solid #336699\">\n  <td class=\"active\"><a class=\"popmake-4415\" href=\"#\">Incorporate Privacy, Civil Liberties, and Security from the Beginning<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3417\" href=\"#\">Use Math to Reduce Bad Outcomes Caused by Math<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3103\" href=\"#\">Promote Better Adoption through Gameplay<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3199\" href=\"#\">Entrust Sector-specific Agencies to Establish AI Standards for Their Domains\u00a0<\/a><\/td>\n <\/tr>\n<\/tbody>\n<\/table>\n<p>[\/et_pb_tab][et_pb_tab title=&#8221;Government Dependence on Black Box Vendors &#8221; _builder_version=&#8221;4.4.8&#8243;]<\/p>\n<div style=\"padding: 0px 0px 15px 0px;margin: 0px 0px 0px 0px\">\n<h3>Government Dependence on Black Box Vendors<\/h3>\n<p>Trade secrecy and proprietary products make it challenging to verify and validate the relevance and accuracy of vendors\u2019 algorithms. These examples demonstrate the importance of at least knowing the attributes of the data and processes for creating the AI model.<\/p>\n<p>&nbsp;<\/p>\n<\/div>\n<div class=\"example-callout\">\n<p><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-4030 size-full alignleft z-index40\" src=\"https:\/\/sites.mitre.org\/aifails\/wp-content\/uploads\/sites\/15\/2020\/03\/icon_example_robot-002.png\" alt=\"\" width=\"100\" height=\"83\" \/><\/p>\n<h3>Examples<\/h3>\n<p>COMPAS, a tool that assesses recidivism risk of prison inmates (repeating or returning to criminal behavior), produced controversial results. In one case, because of an error in the data fed into the AI, an inmate was denied parole despite having a nearly perfect record of rehabilitation. Since COMPAS is proprietary, neither judges nor inmates know how the tool makes its decisions.<a class=\"reference\" href=\".\/references\/#12.1\" target=\"_blank\" rel=\"noopener noreferrer\">1,<\/a><a class=\"reference\" href=\".\/references\/#12.2\" target=\"_blank\" rel=\"noopener noreferrer\">2<\/a><\/p>\n<p>The Houston Independent School District implemented an AI to measure teachers\u2019 performances by comparing their student\u2019s test scores to the statewide average. The teacher\u2019s union won a lawsuit, arguing that the proprietary nature of the product prevents teachers from verifying the results, thereby violating their Fourteenth Amendment rights to due process.<a class=\"reference\" href=\".\/references\/#12.3\" target=\"_blank\" rel=\"noopener noreferrer\">3<\/a><\/p>\n<\/div>\n<h3>Why is this a fail?<\/h3>\n<p>For government organizations, it\u2019s cheaper or easier to acquire algorithms from or outsource algorithm development to third-party vendors. To verify and validate the delivered technology, the government agency needs to understand the methodology that produced it: from analyzing what datasets were applied to knowing the objectives of the AI model to ensuring the operational environment was captured correctly.<\/p>\n<p>&nbsp;<\/p>\n<h3>What happens when things fail?<\/h3>\n<p>Often the problems with the vendors\u2019 models come about because the models\u2019 proprietary nature inhibits verification and validation capabilities. For example, if the vendor modified or added to the training data that the government supplied for the algorithm, or if the government\u2019s datasets and operating environment have evolved from those provided to the vendor, then the AI won\u2019t perform as expected. Unless the contract says otherwise, the vendor keeps its training and validation processes private.<\/p>\n<p>In certain cases the government agency doesn\u2019t have a mature enough understanding of AI requirements and acquisition to prevent mistakes. Sometimes a government agency doesn\u2019t buy a product, but it buys a service. For example, since government agencies usually don\u2019t have fully AI-capable workforces, an agency might provide its data to the vendor with the expectation that the vendor\u2019s experts might discover patterns in the data. In some of these instances, agencies have forgotten to keep some data to serve as a test set, since the same data cannot be used for training and testing the product.<\/p>\n<p>These verification and validation challenges will become more important, yet harder to overcome, as vendors begin to pitch end-to-end AI platforms rather than specialized AI models.<\/p>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<table style=\"border: 5px solid #336699\">\n<tbody>\n <tr>\n  <th class=\"lessonshdr\" colspan=\"4\"><\/th>\n <\/tr>\n <tr>\n  <th class=\"lessons2hdr\" colspan=\"4\"><\/th>\n <\/tr>\n <tr>\n  <th class=\"categoryhdr1\"><\/th>\n  <th class=\"categoryhdr2\"><\/th>\n  <th class=\"categoryhdr3\"><\/th>\n  <th class=\"categoryhdr4\"><\/th>\n <\/tr>\n <tr>\n  <td class=\"active\"><a class=\"popmake-3347\" href=\"#\">Hold AI to a Higher Standard<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3405\" href=\"#\">Involve the Communities Affected by the AI <\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3421\" href=\"#\">Make Our Assumptions Explicit<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-2906\" href=\"#\">Monitor the AI&#8217;s Impact and Establish Layers of Accountability<\/a><\/td>\n <\/tr>\n <tr>\n  <td class=\"active\"><a class=\"popmake-3096\" href=\"#\">It&#8217;s OK to Say No to Automation<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3409\" href=\"#\">Plan to Fail<\/a><\/td>\n  <td class=\"inactive\"><a class=\"popmake-3450\" href=\"#\">Try Human-AI Couples Counseling<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3552\" href=\"#\">Envision Safeguards for AI Advocates<\/a><\/td>\n <\/tr>\n <tr>\n  <td class=\"active\"><a class=\"popmake-3393\" href=\"#\">AI Challenges are Multidisciplinary, so They Require a Multidisciplinary Team<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3414\" href=\"#\">Ask for Help: Hire a Villain<\/a><\/td>\n  <td class=\"inactive\"><a class=\"popmake-4413\" href=\"#\">Offer the User Choices<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3146\" href=\"#\">Require Objective, Third-party Verification and Validation\u00a0<\/a><\/td>\n <\/tr>\n <tr style=\"border-bottom: 5px solid #336699\">\n  <td class=\"inactive\"><a class=\"popmake-4415\" href=\"#\">Incorporate Privacy, Civil Liberties, and Security from the Beginning<\/a><\/td>\n  <td class=\"inactive\"><a class=\"popmake-3417\" href=\"#\">Use Math to Reduce Bad Outcomes Caused by Math<\/a><\/td>\n  <td class=\"inactive\"><a class=\"popmake-3103\" href=\"#\">Promote Better Adoption through Gameplay<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3199\" href=\"#\">Entrust Sector-specific Agencies to Establish AI Standards for Their Domains\u00a0<\/a><\/td>\n <\/tr>\n<\/tbody>\n<\/table>\n<p>[\/et_pb_tab][et_pb_tab title=&#8221;Clear as Mud&#8221; _builder_version=&#8221;4.4.8&#8243;]<\/p>\n<div style=\"padding: 0px 0px 15px 0px;margin: 0px 0px 0px 0px\">\n<h3>Clear as Mud<\/h3>\n<p>The technical and operational challenges in creating a perfectly understandable model can dissuade developers from including incomplete, but still helpful, context and explanations. This omission can prevent people from using an otherwise beneficial AI.<\/p>\n<p>&nbsp;<\/p>\n<\/div>\n<div class=\"example-callout\">\n<p><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-4030 size-full alignleft z-index40\" src=\"https:\/\/sites.mitre.org\/aifails\/wp-content\/uploads\/sites\/15\/2020\/03\/icon_example_robot-002.png\" alt=\"\" width=\"100\" height=\"83\" \/><\/p>\n<h3>Examples<\/h3>\n<p>When UPS rolled out a route-optimization AI that told drivers the best route to take, drivers initially rejected it because they felt they knew better. Once UPS updated the system to provide explanations for some of its suggestions, the program had better success.<a class=\"reference\" href=\".\/references\/#13.1\" target=\"_blank\" rel=\"noopener noreferrer\">1<\/a><\/p>\n<p>A psychiatrist realized that Facebook\u2019s \u2018people you may know\u2019 algorithm was recommending her patients to each other as potential \u2018friends,\u2019 since they were all visiting the same location.<a class=\"reference\" href=\".\/references\/#13.2\" target=\"_blank\" rel=\"noopener noreferrer\">2<\/a> Explanations to both users and developers as to why this algorithm made its recommendations could have mitigated similar breaches of privacy and removed those results from the output.<\/p>\n<\/div>\n<h3>Why is this a fail?<\/h3>\n<p>When we introduce an AI into a new system or process, each set of stakeholders \u2013 AI developers, operators, decision makers, affected communities, and objective third-party evaluators \u2013 has different requirements for understanding, using, and trusting the AI system.<a class=\"reference\" href=\".\/references\/#13.3\" target=\"_blank\" rel=\"noopener noreferrer\">3<\/a> These requirements are also domain and situation specific.<a class=\"reference\" href=\".\/references\/#13.4\" target=\"_blank\" rel=\"noopener noreferrer\">4<\/a><\/p>\n<p>Especially as we begin to develop and adopt AI products that enhance or substitute for human judgment, it is essential that users and policymakers know more about how an AI functions and the intended and non-intended uses for the AI. Adding explanations, documentation, and context are so important because they help calibrate trust in an AI \u2013 that is, <em>figuring out how to trust the AI to the extent it should be trusted<\/em>. Empowering users and stakeholders with understanding can address concepts such as:<\/p>\n<ul>\n<li>Transparency \u2013 how does the AI work and what are its decision criteria?<\/li>\n<li>Traceability \u2013 can the AI help developers and users follow and justify its decision-making process?<\/li>\n<li>Interpretability \u2013 can developers and users understand and make sense of any provided explanations?<\/li>\n<li>Informativeness \u2013 does the AI provide information that different stakeholders find useful?<\/li>\n<li>Policy \u2013 under what conditions is the AI used and how is it incorporated into existing processes or human decision making?<\/li>\n<li>Limitations \u2013 do the stakeholders understand the limits of the AI and its intended uses?<a class=\"reference\" href=\".\/references\/#13.5\" target=\"_blank\" rel=\"noopener noreferrer\">5,<\/a><a class=\"reference\" href=\".\/references\/#13.6\" target=\"_blank\" rel=\"noopener noreferrer\">6,<\/a><a class=\"reference\" href=\".\/references\/#13.7\" target=\"_blank\" rel=\"noopener noreferrer\">7<\/a><\/li>\n<\/ul>\n<p>Traditionally, the conversation in the AI community has focused on transparency (AI experts refer to it as \u201cexplainability\u201d or \u201cexplainable AI\u201d). Approaches for generating AI explanations are very active areas of research, but coming up with useful explanations of how the model actually makes decisions remains challenging for several reasons. Technically, it can be hard because certain models are very complex. Current explainer tools can emphasize which inputs had the most influence on an answer, but not why they had that influence, which makes them valuable but incomplete. Finally, early research showed a tradeoff between accuracy and explainability, but this tradeoff may not always exist. Some of us have responded to the myth that there must be a tradeoff by overlooking more interpretable models in favor of more common but opaque ones.<a class=\"reference\" href=\".\/references\/#13.8\" target=\"_blank\" rel=\"noopener noreferrer\">8<\/a><\/p>\n<p>&nbsp;<\/p>\n<h3>What happens when things fail?<\/h3>\n<p>Cognitively, existing explanations can be misleading. Users can be tempted to impart their own associations or anthropomorphize an AI (i.e., attributing human intentions to it). Also, assuming causality when there is only correlation in an AI system will lead to incorrect conclusions.<a class=\"reference\" href=\".\/references\/#13.9\" target=\"_blank\" rel=\"noopener noreferrer\">9<\/a> If these misunderstandings can cause financial, psychological, physical, or other types of harm, then the importance of good explanations becomes even greater.<a class=\"reference\" href=\".\/references\/#13.a\" target=\"_blank\" rel=\"noopener noreferrer\">10<\/a><\/p>\n<blockquote>\n<p>Adding explanations, documentation, and context are so important because they help calibrate trust in an AI \u2013 that is, figuring out how to trust the AI to the extent it should be trusted<\/p>\n<\/blockquote>\n<p>The challenge lies in expanding the conversation beyond transparency and explainability to include the multitude of ways in which AI stakeholders can improve their understanding and choice. If we adopt the mindset that the users, policymakers, auditors, and others in the AI workflow are all our customers, this can help us devote more resources to providing the context that these stakeholders need.<\/p>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<table style=\"border: 5px solid #336699\">\n<tbody>\n <tr>\n  <th class=\"lessonshdr\" colspan=\"4\"><\/th>\n <\/tr>\n <tr>\n  <th class=\"lessons2hdr\" colspan=\"4\"><\/th>\n <\/tr>\n <tr>\n  <th class=\"categoryhdr1\"><\/th>\n  <th class=\"categoryhdr2\"><\/th>\n  <th class=\"categoryhdr3\"><\/th>\n  <th class=\"categoryhdr4\"><\/th>\n <\/tr>\n <tr>\n  <td class=\"active\"><a class=\"popmake-3347\" href=\"#\">Hold AI to a Higher Standard<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3405\" href=\"#\">Involve the Communities Affected by the AI <\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3421\" href=\"#\">Make Our Assumptions Explicit<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-2906\" href=\"#\">Monitor the AI&#8217;s Impact and Establish Layers of Accountability<\/a><\/td>\n <\/tr>\n <tr>\n  <td class=\"inactive\"><a class=\"popmake-3096\" href=\"#\">It&#8217;s OK to Say No to Automation<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3409\" href=\"#\">Plan to Fail<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3450\" href=\"#\">Try Human-AI Couples Counseling<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3552\" href=\"#\">Envision Safeguards for AI Advocates<\/a><\/td>\n <\/tr>\n <tr>\n  <td class=\"active\"><a class=\"popmake-3393\" href=\"#\">AI Challenges are Multidisciplinary, so They Require a Multidisciplinary Team<\/a><\/td>\n  <td class=\"inactive\"><a class=\"popmake-3414\" href=\"#\">Ask for Help: Hire a Villain<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-4413\" href=\"#\">Offer the User Choices<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3146\" href=\"#\">Require Objective, Third-party Verification and Validation\u00a0<\/a><\/td>\n <\/tr>\n <tr style=\"border-bottom: 5px solid #336699\">\n  <td class=\"active\"><a class=\"popmake-4415\" href=\"#\">Incorporate Privacy, Civil Liberties, and Security from the Beginning<\/a><\/td>\n  <td class=\"inactive\"><a class=\"popmake-3417\" href=\"#\">Use Math to Reduce Bad Outcomes Caused by Math<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3103\" href=\"#\">Promote Better Adoption through Gameplay<\/a><\/td>\n  <td class=\"active\"><a class=\"popmake-3199\" href=\"#\">Entrust Sector-specific Agencies to Establish AI Standards for Their Domains\u00a0<\/a><\/td>\n <\/tr>\n<\/tbody>\n<\/table>\n<p>[\/et_pb_tab][\/et_pb_tabs][\/et_pb_column_inner][\/et_pb_row_inner][\/et_pb_column][\/et_pb_section][et_pb_section fb_built=&#8221;1&#8243; fullwidth=&#8221;on&#8221; disabled_on=&#8221;off|off|off&#8221; _builder_version=&#8221;4.4.8&#8243; module_alignment=&#8221;center&#8221; global_module=&#8221;3880&#8243; saved_tabs=&#8221;all&#8221;][et_pb_fullwidth_code disabled_on=&#8221;off|off|off&#8221; admin_label=&#8221;Footer menu&#8221; _builder_version=&#8221;4.4.8&#8243; background_color=&#8221;#d5dde0&#8243; text_orientation=&#8221;center&#8221; module_alignment=&#8221;center&#8221; custom_padding=&#8221;10px||10px||false|false&#8221;]Add Your Experience! This site should be a community resource and would benefit from the addition of other examples and voices. You can write to us by clicking <a href=\"mailto:jrotner@mitre.org;rhodge@mitre.org;ldanley@mitre.org?subject=AI Fails website\">here<\/a>.[\/et_pb_fullwidth_code][\/et_pb_section]<\/p>\n","protected":false},"excerpt":{"rendered":"<p>We&#8217;re Not Done Yet: After Developing the AI Developing AI is a dynamic, multifaceted process. Even if an AI performs optimally from a technical standpoint, other constraining factors could limit its overall performance and acceptance. Developing an AI to be safe and dependable means stakeholders must learn more about how the AI functions as the [&hellip;]<\/p>\n","protected":false},"author":142,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"_acf_changed":false,"_et_pb_use_builder":"on","_et_pb_old_content":"","_et_gb_content_width":"","footnotes":""},"class_list":["post-58","page","type-page","status-publish","hentry"],"acf":[],"_links":{"self":[{"href":"https:\/\/sites.mitre.org\/aifails\/wp-json\/wp\/v2\/pages\/58","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/sites.mitre.org\/aifails\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/sites.mitre.org\/aifails\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/sites.mitre.org\/aifails\/wp-json\/wp\/v2\/users\/142"}],"replies":[{"embeddable":true,"href":"https:\/\/sites.mitre.org\/aifails\/wp-json\/wp\/v2\/comments?post=58"}],"version-history":[{"count":0,"href":"https:\/\/sites.mitre.org\/aifails\/wp-json\/wp\/v2\/pages\/58\/revisions"}],"wp:attachment":[{"href":"https:\/\/sites.mitre.org\/aifails\/wp-json\/wp\/v2\/media?parent=58"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}