building-ai-trust-the-key-role-of-explainability
building-ai-trust-the-key-role-of-explainability
November 2024
Artificial intelligence has the potential to Why invest in this capability:
deliver massive gains in economic productivity Getting ROI from XAI
and enable positive social change around the As with any investment in an uncertain environment,
world. So it’s little surprise that the number of organizations seeking to enhance AI explainability
companies adopting AI-powered software, tools, must consider the benefits and costs to decide
and platforms, including generative AI (gen AI), has how and when to act in the absence of perfect
surged throughout 2024. But that enthusiasm has information on the potential upside and risks
been accompanied by a fair amount of trepidation: involved. Today’s AI landscape is fraught with
in McKinsey research, 91 percent of respondents uncertainty, and in this context, leading AI labs like
doubt their organizations are “very prepared” Anthropic are making bets that investments in XAI
to implement and scale the technology safely will pay off as a path to differentiation in a crowded
and responsibly.1 Such doubt is understandable. field of foundation model builders (see sidebar
Along with its potential to boost productivity “The evolution of XAI and today’s challenges”).
and innovation, gen AI in particular poses novel Meanwhile, enterprises are seeking to meet the
risks—for example, hallucinations and inaccurate or expectations of their stakeholders and regulators.
biased outputs—which threaten to undermine trust
in the technology. One thing is certain: demand for XAI is rising. As
global AI regulations begin to take shape, the need
To capture the full potential value of AI, for explainability and interpretation is increasing,
organizations need to build trust. Trust, in fact, with more organizations seeking guidelines on
is the foundation for adoption of AI-powered how to determine what level of explainability to
products and services. After all, if customers adopt and how much information to release about
or employees lack trust in the outputs of AI their models. The EU AI Act, for example, imposes
systems, they won’t use them. Trust in AI comes specific transparency requirements for different
via understanding the outputs of AI-powered AI use cases classified according to its risk-based
software and how—at least at a high level—they framework. For example, in the case of high-risk
are created. Organizations increasingly recognize AI systems—such as systems used in recruitment,
this. In a McKinsey survey of the state of AI in 2024, like résumé-ranking software—organizations are
40 percent of respondents identified explainability required to provide information about the system’s
as a key risk in adopting gen AI. Yet at the same capabilities, limitations, data lineage, and the logic
time, only 17 percent said they were currently behind the decisions it makes.
working to mitigate it.2
Imagine driving a car. Setting a speed limit of
This conundrum has raised the need for enhanced 45 miles per hour is useless if your vehicle lacks
AI explainability (XAI)—an emerging approach to a speedometer that indicates where you are
building AI systems designed to help organizations relative to the standard. Similarly, to respond
understand the inner workings of those systems to AI regulations, organizations need methods
and monitor the objectivity and accuracy of their that provide visibility into how AI models are
outputs. By shedding some light on the complexity built and how they can be tested before release.
of so-called black-box AI algorithms, XAI can Organizations also need observability solutions
increase trust and engagement among those that provide sufficient insight into their AI models
who use AI tools. This is an essential step as to ensure they comply with regulations and their
AI initiatives make the difficult journey from own values and standards. This raises crucial
early use case deployments to scaled, enterprise- questions: Are organizations prepared to deliver this
wide adoption. level of transparency? Do they have the necessary
capabilities and technologies in place? Have
1
“Implementing generative AI with speed and safety,” McKinsey Quarterly, March 13, 2024.
2
“The state of AI in early 2024: Gen AI adoption spikes and starts to generate value,” McKinsey, May 30, 2024.
The field of AI explainability has evolved example, a ProPublica investigation into for example, has provided significant
significantly in recent years. Early AI the COMPAS algorithm, used by US courts improvements to techniques for LLM
tools, employing rule-based systems to assess the likelihood of a defendant explainability and interpretability. Tools
and decision trees, were relatively simple reoffending, revealed systematic bias to interpret the behavior of language
and transparent by design. However, as against African American defendants. models, including OpenAI’s transformer
machine learning models have grown Unfortunately, addressing these concerns debugger, are new and only beginning to be
more complex, it has become more is no simple matter. One major issue is the understood and implemented. An example
difficult to trace the reasons underpinning increasing complexity of advanced large of how tech companies are incorporating
their decision-making processes. The language models (LLMs), which rely on explainability tools into their platforms
early 2000s saw the development of deep neural networks and often operate is Google’s Vertex Explainable AI, which
methods like local interpretable model- as black boxes, with opaque decision- enhances understanding of generative
agnostic explanations (LIME) and making processes. And the lack of access AI and LLMs through feature-based and
Shapley additive explanations (SHAP), to the architecture of proprietary models example-based explanations that give
which provided insights into individual makes it difficult to understand how they users insights into model predictions by
predictions of complex models. Google operate. Previously, teams could control identifying influential features in complex
introduced its What-If Tool, enhancing fairness by curating training data and generative models like transformer-based
model transparency through interactive applying calibration techniques. However, LLMs. In addition, recent community-driven
visualizations; IBM released the AI today’s LLMs resist such control, making research, like work on behavior analysis
Explainability 360 tool kit; and DARPA explainability difficult. Finally, organizations at the head level of LLM architectures,
produced an Explainable AI (XAI) program, increasingly face a trade-off between reflects growing momentum toward
which further advanced the field by model accuracy and interpretability: more unpacking model behaviors. The scale and
developing comprehensive tool kits and complex models and LLMs often achieve complexity of more mature techniques for
techniques to interpret AI models. higher accuracy but at the cost of being unpacking these intricate systems present
less interpretable and harder to explain. unprecedented challenges, but even if
In the meantime, several highly publicized
much work remains, we anticipate progress
missteps have highlighted the growing Innovators are focused on these issues, and
in the coming years.
need for AI explainability. In 2016, for some strides have been made. Anthropic,
platforms and innovators created reliable methods into the SDLC right from the conception of new
of measurement? AI-powered offerings. These experts can form an
XAI center of excellence (COE) to provide expertise
XAI is best thought of as a set of tools and practices and training across teams, reshaping the software
designed to help humans understand why an AI development life cycle and assuring coordinated
model makes a certain prediction or generates enterprise-wide investments in tools and training.
a specific piece of content. The ultimate goal is The COE also can address the need for additional
to ensure that these outputs are of high quality, compute power and cloud consumption to deliver
untainted by bias, inaccuracy, or hallucination. the additional training, post-training, and production
This requires several kinds of investment—in tools, monitoring essential to enhancing explainability.
people, and processes. Building more explainable AI
and machine learning solutions requires deployment How can we ensure a return on these investments
of new technology in the software delivery life cycle in technologies that are often in their early stages?
(SDLC) from the start, when models are trained While XAI is still emerging from research-focused
and tested or as pretrained models are fine-tuned, efforts in academia and R&D labs into real-world
ending when code moves into production and applications, its benefits are far more tangible than
ongoing monitoring and observability are needed. commonly thought. We see five areas where XAI
Expertise in XAI techniques must be built via hiring can deliver a return that drives positive ROI:
and/or training, and the experts must be integrated
One of the advantages of XAI is that make AI’s complex logic more accessible— understanding through a shared language
it places humans at the center of AI and therefore truly useful.” of curiosity and exploration,” she says.
efforts. “Data storytelling plays a crucial “It’s not just about demystifying AI; it’s
According to Lupi, explainability efforts
role in bridging the gap between human about finding poetry in how we learn to
are fundamentally about humanizing the
understanding and AI,” says Giorgia Lupi, work alongside it. Think of AI like another
machine’s inner workings and framing AI’s
a partner at the design firm Pentagram and member of the team that is from another
data as stories that reveal its logic. “When
creator of the Data-Humanism Manifesto. culture. We, as humans, need to be open
we embrace storytelling to articulate how
“By translating the machine’s thought to learning how AI reasons, even if it’s not
AI ‘thinks,’ we invite people to connect
processes into narratives that resonate immediately intuitive for us.”
with this new kind of teammate, fostering
with our natural ways of learning, we can
developed by researchers and engineers to respond — Developers need explanations of the models’
to the needs and questions of a diverse group of functioning so they can improve and debug
stakeholders and users. This emerging talent will be these nondeterministic systems, add post-
the key to designing XAI that works for all. training enhancements, and ensure the AI
models deliver expected outcomes.
Stakeholder needs can be broken down into
six personas, each benefitting from different Beyond these different stakeholders, varying
techniques and explanations: contexts and risk scenarios influence the format of
the explanations provided. Explanations can take the
— Executive decision makers require enough form of data visualizations or text reports and will vary
understanding and information about models in technical detail. Understanding the specific needs
to be accountable for their actions with respect of each stakeholder at a particular time is essential
to customers and employees—specifically, to to providing effective and meaningful AI explanations
ensure that models behave in alignment with that meet their unique needs.
the organization’s strategies, brand ethos,
and values.
How does XAI work, and what
— AI governance leaders constitute a cross- techniques are available?
functional group—drawn from functions like To meet these diverse needs, the XAI community
legal, risk, information security, engineering, continues to create new explainability techniques,
and product—that is responsible for shaping AI which involve algorithms to make the decision-
systems in accordance with policies, standards, making processes of AI models more transparent.
and regulations. These can be grouped based on stakeholders’
intents and goals. In general, techniques can
— Affected users need explanations about the be categorized along two dimensions: when the
outcomes they get from AI models. explanation is produced (before or after the model
is trained) and the scope of the explanation (global
— Business users require insights to enhance or local).
everyday decision making, improve processes,
and optimize operational efficiency. The first macro category of XAI techniques
comprises “post-hoc methods,” which involve
— Regulators/auditors require explanations analyzing models after they have been trained,
and interpretability from models to make in contrast to “ante-hoc methods,” which refer
sure they are safe and compliant as rules and to intrinsically explainable models, like decision
regulations evolve. trees. For example, when an HR department
Carlo Giovine is a partner in McKinsey’s London office, where Mara Pometti is a consultant; Roger Roberts is a partner in the
Bay Area office, where Medha Bankhwal is an associate partner.
This article was edited by Larry Kanter, a senior editor in the New York office.