THE age of pervasive AI is here.1 Since 2017, Deloitte’s annual State of AI in the Enterprise report has measured the rapid advancement of AI technology globally and across industries. In the most recent edition, published in July 2020, a majority of those surveyed reported significant increases in AI investments, with more than three-quarters believing that AI will substantially transform their organization in the next three years. In addition, AI investments are increasingly leading to measurable organizational benefits: improved process efficiency, better decision-making, increased worker productivity, and enhanced products and services.2 These possible benefits have likely driven the growth in AI’s perceived value to organizations—nearly three-quarters of respondents report that AI is strategically important, an increase of 10 percentage points from the previous survey.
However, a growing unease threatens this rising trendline: Fifty-six percent of surveyed organizations say they plan to slow or are already slowing AI adoptions because of concern about emerging risks—a remarkable level of apprehension considering AI’s acknowledged benefits and strategic importance. To better understand the underlying issues at stake, we analyzed respondents’ concerns based on three main categories: confidence in the AI decision-making process, ethics of AI and data use, and marketplace uncertainties.3
We aim to explore these concerns in detail and offer insight into what companies can do to competently manage these risks. By doing so, organizations can build the internal assuredness needed to continue investing and innovating—and increase external stakeholders’ trust that AI strategies can be executed as intended.
Discussion of AI-related risks has been growing in recent years for both business leaders and the general public. Debates about whether AI will kill or create jobs are ongoing; meanwhile, reports of AI bias or failures frequently break into the headlines.4 In July 2020, public backlash and widespread criticism shut down AI-powered startup Genderify after only five days. The company claimed that its AI could identify a person’s gender by analyzing their name, username, and email address; it pitched this capability to businesses as an enhancement to customer data. The system reportedly included various gender misconceptions and made decisions based on faulty assumptions. For instance, records containing “scientist” returned only a 4.3% probability that the individual was female, despite far higher current representation of women in the sciences.5 “Professor” resulted in a 98.4% probability for male, when men hold only 74% of tenured positions in the United States (a percentage that itself reflects inequality).6
Examples such as this—and the public backlash that often accompanies their exposure—are making many leaders understandably skittish about how they roll out AI within their organization and for customers, employees, and wider ecosystem partners. To understand more precisely where organizations’ biggest apprehensions originate, we asked survey respondents to rate a list of concerns on a scale from “minimal” to “extreme.” We grouped the listed risks into three categories: confidence, ethics, and marketplace uncertainties.
Confidence refers to whether the company generally believes an AI tool itself is reliable. This includes trusting the insights and decisions made by AI, trust in system security, and being able to understand, justify, or explain AI’s decision-making process. Ethics refers to questioning whether using AI is good and right for society, regardless of AI’s benefit to the business, including issues of data privacy, fairness, bias, and the potential for job loss. Marketplace uncertainties are factors that are further outside the company’s direct control, such as the changing regulatory landscape and public or employee opinion. These factors are not necessarily related to the quality of the AI, but they affect how organizations can implement or use it.
Figure 1 reflects the percentages of respondents who rated two or more concerns in each category as “major” or “extreme.”
As a category, confidence in AI systems was the most significant, with 73% of all business leaders surveyed identifying two or more concerns as major or extreme. When looking specifically at the respondents planning to slow their AI adoption, 85% claim two or more major or extreme confidence-related risks.
It’s understandable that concerns around confidence could lead to apprehension. They are most tightly related to the quality of the products and services a company provides and thus can contribute substantially to an organization’s reputation and public perception; they also can affect the very deliverables a company puts out into the world for consumption. For example, a medical provider could use AI’s immense potential to identify cancer types and determine optimal treatments.7 It is possible that AI decision engines could recommend unsafe treatments or misidentify types of cancer. This can result in serious consequences for patients, expose companies to risk, and result in the loss of AI investment dollars. Unsuccessful initiatives can lead to damages more costly than just lost investment—they can reduce the confidence of an organization’s employees in further innovation and, worse, reduce the trust of patients or customers.
In the case above, the company reportedly created safeguards to ensure patient safety, reducing the likelihood that it would damage patients’ trust. Additionally, the mistakes were used to identify system failures and improve future algorithmic solutions.8 If mitigated well from the beginning, this can be turned into a positive element of the path to AI maturity. In fact, there is evidence suggesting that as companies mature in their AI capabilities, their level of concern tends to shift as they develop processes, behaviors, or skills to lower their risks and drive more positive outcomes (see sidebar, “AI maturity and how organizations learn to manage perceived risks”).
Across the board, as surveyed companies mature in their AI capabilities, the level of concern across confidence, ethics, and marketplace uncertainties follows a relatively bell-shaped trajectory.
Both low- and high-maturity organizations that were surveyed report lower levels of concern, whereas medium-maturity organizations report the highest level of concern across all three categories. This may be because low-maturity organizations don’t yet have full vantage of the risks, with many projects still in proof-of-concept or pilot mode. Then, as organizations achieve a medium level of maturity, the challenges might become more apparent, but they may not have developed the capabilities needed to address them yet. Finally, as organizations reach an advanced level, they could have more of the capabilities needed to mitigate those risks, and so their level of concern often decreases once again.
Another significant challenge related to AI confidence is that today’s AI systems tend to come with a tradeoff between interpretability and power. Some machine-learning models are so complex that even highly trained data scientists have difficulty understanding precisely how the algorithms make decisions. As the use of neural networks rises, this issue is becoming even more pronounced, further implicating organizations’ ability to justify decisions, mitigate errors, and satisfy regulators.
To help solve this, explainable AI is a developing computer science field that seeks to create AI models that are better able to explain themselves. At present, explainable AI can provide general information about how an AI program makes a decision by disclosing that program’s strengths and weaknesses and the specific criteria the program used to arrive at a decision, and advise on appropriate levels of trust in various types of decisions. Recent recommendations from the U.S. Department of Commerce suggest four principles for assessing AI’s explainability: AI systems need to provide accompanying evidence or reasons for all outputs; systems must provide meaningful and understandable explanations to individual users; system explanations must correctly reflect the system’s process for generating the output; and the system should operate only under the conditions for which it was designed and should not supply decisions to a user without sufficient confidence.9 Mastering AI explainability should, in turn, influence an organization’s ability to use AI ethically.
Fifty-seven percent of respondents reported two or more ethics-related concerns as major or extreme. Looking specifically at the respondents planning to slow their AI adoption, 73% cited a major or extreme concern about in at least two ethical areas.
A unique challenge with AI ethics, especially around fairness and bias, is that the data upon which the AI is built can itself be incomplete, biased, or unequal. For this reason, it can be difficult to root out all the unintended biases within a data set, even when a company is well-intentioned.
The humans building AI, analyzing its outputs, and applying its solutions can also fall prey to unintended biases, requiring deliberate actions to mitigate these wherever possible. In a recent health care example, a risk-prediction algorithm using health care spending as a proxy for care ultimately demonstrated racial bias in its results, giving white patients a better chance at benefiting from an extra care program than Black patients.10 While the algorithm accurately reflected health spending of both groups, it failed to account for economic disadvantages Black people tend to face that often affect their health care spending: more expensive emergency room visits, lower insurance coverages, etc.11 This could result in similar spending levels for vastly different levels of need—a reality that the system’s analysis of proxy data missed. In the above example, Black patients were given incorrect risk scores and were excluded from extra care programs at a greater rate.
It is important that program designers thoroughly investigate proxy decisions and variable creation, with safeguards for evaluating whether that use reflects the lived realities of your stakeholders. Further, such examples underscore the need for diverse representation, as well as skills and training among engineers and decision-makers around possible unintended outcomes of algorithm creation.
Finally, 55% of respondents reported two or more marketplace uncertainties as major or extreme concerns. Despite ranking third as a category, this area of concern seems to have an outsized link to investment behaviors: Of the respondents expecting to slow AI adoption, 71% reported having at least two significant marketplace uncertainties.
The concerns in this category are wide-ranging, from public opinion to the ambiguous regulatory landscape. Fifty-seven percent of survey respondents cite major or extreme concerns about new and changing rules and regulations of AI technologies, such as data privacy, facial recognition, and decision transparency.12 Fifty-two percent cite major or extreme concerns about customer backlash should they find a flaw or privacy violation in an AI application. Just over half also worry about negative employee perceptions when AI systems are used.
The immense degree of transformation and the vast potential for risk that AI offers can seem daunting—only four in 10 of our survey respondents claimed full preparedness.13 Given AI technologies’ rapid development, a natural level of uncertainty is sure to linger for some time. However, organizations can look to some current frameworks—and lessons learned from previous technology disruptions—for help in navigating today’s uncertainties.
Of the three categories of concern analyzed previously, two of them—confidence and ethics—land primarily within the organization’s locus of control. They both ultimately point to the question of whether an organization’s use of AI is trustworthy. To increase confidence in AI, organizations should ensure that their tools and solutions are transparent, reliable, and safe, and that there is a system of accountability in place. Reducing concerns around ethical use will often require evaluating whether tools and solutions are designed for fairness—both in intent and in their effect—and that data use follows clear privacy standards. To help organizations frame AI risks, leaders should develop safeguards that include addressing the six key dimensions highlighted below in figure 6.
Confidence and ethics concerns essentially come down to a question of whether an organization’s use of AI is trustworthy. An organization’s ability to implement this framework largely depends on responsible data management, strong governance standards, and ensuring that a variety of perspectives are in the room to identify and speak out against harmful assumptions.
Recommendation: Focus on data management and governance. An organization’s ability to skillfully work with data is critical to AI quality and explainability, increasing confidence in AI and leaders’ preparedness to manage its ethical implications. Even today, 40% of surveyed organizations still report “low” or “medium” levels of sophistication across a range of data practices, and nearly a third of executives identified data-related challenges among the top three concerns hampering their company’s AI initiatives.14 When it comes to governance, only one in five companies surveyed routinely monitor, manage, and improve data quality as part of a formal data governance effort, while only 12% of organizations trust their data to be up to date, and mere 9% believe the data is accurate.15
Without skillful data management and governance processes in place, an organization could struggle to mitigate risks. By focusing on these two foundational elements, companies can put themselves in a position to successfully implement all six dimensions of trustworthy AI and address the confidence and ethics concerns that can threaten to slow their adoption.
Recommendation: Insist on diversity of thought. A company’s ability to implement safeguards also depends on a diversity of perspectives within technology and business stakeholder groups. Data capabilities and governance processes create checkpoints to evaluate compliance on a variety of standards and regulations, but without a diversity of perspectives, important insights may not surface. Ensuring diversity of thought can help increase the likelihood that potential problems are flagged during the appropriate design and review phases—and not after product launch.
Managing marketplace uncertainties, those concerns that relate to more external ambiguities, will likely require different approaches. Changing regulations and fluctuating public and employee opinion are always a feature of new and disruptive technology transformations. While companies can’t always be in complete control of these, leaders can take steps to navigate them more successfully.
Recommendation: Understand the importance of change management and communication. Public and employee opinion often has little to do with how well AI performs. Thoughtful change management and communication can be critical to successfully navigating this uncertainty. AI will likely transform many organizational roles in the coming years, and helping workers and customers learn new ways of working and engaging will be important to success. An example of an implementation is at Humana, which deployed AI customer assistance agents to answer many of the million-plus calls that often overwhelm customer service agents every month. Humana developed AI agents to handle basic information requests, which accounted for 60% of call volume. This isn’t the only way that AI robots equipped with natural language understanding help—they can also assist human agents to gather information behind the scenes, which assisted the Humana call center associates in providing information. By positioning this new technology as an assistant to human workers, it mitigated the often-perceived threat that some employees may feel toward AI. The company focused on using AI to support (not replace) employees on communication, for example, appropriate expressions of empathy, and on how to collaborate with AI to solve problems. These efforts were focused on trying to maximize the possibilities for success.16
Understanding employee perceptions and planning technology rollouts that mitigate their disruptive nature are often critical for buy-in. Consider working with communication and marketing teams to ensure that both employees and customers adequately understand and see value in the way AI is rolling out. Employing a user-centric approach—instead of forcing new behaviors for cost savings—can help to avoid backlash.
Recommendation: Anticipate the capabilities needed to respond to regulatory shifts. When the European Parliament and European Union created the General Data Protection Regulation (GDPR) in 2016, companies had two years to understand and comply with the regulations before they were implemented in 2018. In the months before the GDPR went into effect, industries that had a history of managing regulations, such as finance, found themselves to be comparatively well prepared.17 These companies had already built skills and operational capabilities to respond to the required changes related to compliance. As discussions of new AI policy and potential technology regulations increase, now is the time to gain these capabilities so that your organization is ready to both participate in the creation of regulations and respond when they are enacted.
AI holds enormously transformative potential. More than 90% of technology executives surveyed agree that AI will be at the center of the next technological revolution, and that it will take on mundane tasks, allowing people to have greater freedom to pursue more creative work and play.18
AI’s potential value to our world could be too high for business leaders to shrink away due to an overabundance of cautions. With thoughtful development of capabilities and processes, leaders can mitigate the risks and challenges that so many fear. Taking these steps now could set up your organization for significant competitive advantage in the future.
This analysis is based on data collected from 2,737 IT and line-of-business executives between October and December 2019. Nine countries were represented, and all participating companies have adopted AI technologies. Respondents were required to meet one of the following criteria: determine AI technology spending and/or approve AI investments; develop AI technology strategies; manage or oversee AI technology implementation; serve as an AI technology subject matter expert; or make or influence decisions around AI technology.
Forty-seven percent were IT executives, with the rest line-of-business executives. Seventy percent were C-level executives: CEOs, presidents, and owners (35%); CIOs and CTOs (32%); and other C-level executives (3%).
To understand more precisely where organizations’ biggest apprehensions originate, we asked survey respondents to rate a list of concerns on a scale from “minimal” to “extreme.” We grouped the listed risks into three categories: confidence, ethics, and marketplace uncertainties.
The Deloitte AI Institute helps organizations transform with AI through cutting-edge research and innovation by bringing together the brightest minds in AI to help advance human-machine collaboration in the Age of With. The Institute was established to advance the conversation and development of AI in order to challenge the status quo. The Deloitte AI Institute collaborates with an ecosystem of industry thought leaders, academic luminaries, startups, research and development groups, entrepreneurs, investors, and innovators. This network, combined with Deloitte’s depth of applied AI experience, can help organizations transform with AI. The Institute will cover a broad spectrum of AI focus areas, with current research on ethics, innovation, global advancements, the future of work, and AI case studies.Learn more