没有合适的资源?快使用搜索试试~ 我知道了~
首页《可解释人工智能》概述
资源详情
资源评论
资源推荐

DARPA’s Explainable
Artificial Intelligence Program
David Gunning, David W. Aha
n Dramatic success in machine learning
has led to a new wave of AI a pplications (for
example, transportation, security, medicine,
finance, defense) that offer tremendous
benefits but cannot explain their decisions
and actions to human users. DARPA’s
explainable artificial intelligence (XAI)
program endeavors to create AI systems
whose learned models and decisions
can be understood and appropriately
trusted by end users. Realizing this goal
requires methods for learning more
explainable models, designing effective
explanation interfaces, and understanding
the psychologic requirements for effective
explanations. The XAI developer teams are
addressing the first two challenges by
creating ML techniques and developing
principles, strategies, and human-computer
interaction techniques for generating effec-
tive explanations. Another XAI team is
addressing the third challenge by summa-
rizing, extending, and applying psychologic
theories of explanation to help the XAI
evaluator define a suitable evaluation
framework, which the developer teams
will use to test their systems. The XAI
teams completed the first of this 4-year
program in May 2018. In a series of
ongoing evaluations, the developer
teams are assessing how well their XAM
systems’ explanations improve user un-
derstanding, user trust, and user task
performance.
A
dvances in machine learning (ML) techniques promise
to produce AI systems that perceive, learn, decide, and
act on their own. However, they will be unable to
explain their decisions and actions to human users. This lack
is especially important for the Department of Defense, whose
challenges require developing more intelligent, autonomous,
and symbiotic systems. Explainable AI will be essential if
users are to understand, appropriately trust, and effectively
manage these artificially intelligent partners. To address this,
DARPA launched its explainable artificial intelligence (XAI)
program in May 2017. DARPA defines explainable AI as AI
systems that can explain their rationale to a human user,
characterize their strengths and weaknesses, and convey an
understanding of how they will behave in the future. Naming
this program explainable AI (rather than interpretable,
comprehensible, or transparent AI, for example) reflects
DARPA’s objective to create more human-understandable AI
systems through the use of effective explanations. It also
reflects the XAI team’s interest in the human psychology of
explanation, which draws on the vast body of research and
expertise in the social sciences.
44 AI MAGAZINE
Copyright © 2019, Association for the Advancement of Artificial Intelligence. All rights reserved. ISSN 0738-4602
Deep Learning and Security

Early AI systems were predominantly logical and
symbolic; they performed some form of logical in-
ference and could provide a trace of their inference
steps, which became the basis for explanation. There
was substantial work on making these systems more
explainable, but they fell short of user needs for
comprehension (for example, simply summarizing
the inner workings of a system does not yield a suf-
ficient explanation) and proved too brittle against
real-world complexities.
Recent AI success is due largely to new ML tech-
niques that construct models in their internal repre-
sentations. These include support vector machines,
random forests, probabilistic graphical models, re-
inforcement learning (RL), and deep learning (DL)
neural networks. Although these models exhibit high
performance, they are opaque. As their use has in-
creased, so has research on explainability from the
perspectives of ML (Chakraborty et al. 2017; Ras et al.
2018) and cognitive psychology (Miller 2017). Simi-
larly, many XAI-related workshops have been held re-
cently on ML (for example, the International Conference
on Machine Learning, the Conference on Neural In-
formation Processing Systems), AI (for example, the In-
ternational Joint Conference on Artificial Intelligence),
and HCI (for example, the Conference on Human-
Computer Interaction, Intelligent User Interfaces) con-
ferences, as have special topic meetings related to XAI.
There seems to be an inherent tension between ML
performance (for example, predictive accuracy) and
explainability; often the highest-performing methods
(for example, DL) are the least explainable, and the
most explainable (for example, decision trees) are the
least accurate. Figure 1 illustrates this with a notional
graph of the performance-explainability trade-off for
various ML techniques.
When DARPA formulated the XAI program, it envi-
sioned three broad strategies to improve explainability,
while maintaining a high level of learning performance,
based on promising research at the time (figure 2): deep
explanation, interpretable models, and model induction.
Deep explanation refers to modified or hybrid DL
techniques that learn more explainable features or
representations or that include explanation genera-
tion facilities. Several design choices might produce
more explainable representations (for example, training
data selection, architectural l ayers, loss functions,
regularization, optimization techniques, training
sequences). Researchers have used deconvolutional
networks to visualize convolutional network layers, and
techniques existed for associating semantic concepts
with deep network nodes. Approaches for generating
image captions could be extended to train a second deep
network that generates explanations without explicitly
identifying the original network’s semantic features.
Interpretable models are ML techniques that learn
more structured, interpretable, or causal models. Early
examples included Bayesian rule lists (Letham et al.
2015), Bayesian program learning, learning models of
causal relationships, and use of stochastic grammars
to learn more interpretable structure.
Model induction refers to techniques that experi-
ment with any given ML model— such as a black
box— to infer an approximate explainable model. For
example, the model-agnostic explanation system of
Ribeiro et al. (2016) inferred explanations by ob-
serving and analyzing the input-output behavior of a
black box model.
DARPA used these strategies to categorize a portfolio
of new ML techniques and provide future practitioners
with a wider range of design options covering the
performance-explainability trade space.
XAI Concept and Approach
The XAI program’s goal is to create a suite of new or
modified ML techniques that produce explainable
models that, when combined with effective explanation
techniques, enable end users to understand, appropri-
ately trust, and effectively manage the emerging gen-
eration of AI systems. The target of XAI is an end user
who depends on decisions or recommendations pro-
duced by an AI system, or actions taken by it, and
therefore needs to understand the system’s rationale.
For example, an intelligence analyst who receives rec-
ommendations from a big data analytics system needs
to understand why it recommended certain activity for
further investigation. Similarly, an operator who tasks
an autonomous vehicle to drive a route needs to un-
derstand the system’s decision-making model to ap-
propriately use it in future missions. Figure 3 illustrates
the XAI concept: provide users with explanations that
enable them to understand the system’s overall
strengths and weaknesses, convey an understanding of
how it will behave in future or different situations, and
perhaps permit users to correct the system’smistakes.
This user-centered concept poses interrelated re-
search challenges: (1) how to produce more explain-
able models, (2) how to design explanation interfaces,
and (3) how to understand the psychologic require-
ments for effective explanations. The first two chal-
lenges are being addressed by the 11 XAI research
teams, which are developing new ML techniques to
produce explainable models, and new principles,
strategies, and HCI techniques (for example, visuali-
zation, language understanding, language generation)
to generate effective explanations. The third challenge
is the focus of another XAI research team that is
summarizing, extending, and applying psychologic
theories of explanation.
The XAI program addresses two operationally rel-
evant challenge problem areas (figure 4): data analytics
(classification of events of interest in heteroge neous
multimedia data) and autonomy (decision policies
for autonomous systems). These areas represent
two important ML problem categories (supervised
learning and RL) and Department of Defense in-
terests (intelligence analysis and autonomous systems).
The data analytics challenge was motivated by a
common problem: intelligence analysts are presented
with decisions and recommendations from big data
SUMMER 2019 45
Deep Learning and Security

analytics algorithms and must decide which to report as
supporting evidence in their analyses and which to
pursue further. These algorithms often produce fal s e
alarms that must be pruned and are subject to concept
drift. Furthermore, these algorithms often make recom-
mendations that the analyst must assess to determine
whether the evidence supports or contradicts their hy-
potheses. Effective explanations will help confront these
issues.
The autonomy challenge was motivated by the
need to effectively manage AI partners. For example,
the Department of Defense seeks semiautonomous
systems to augment warfighter capabilities. Operators
will need to understand how these behave so they can
determine how and when to best use them in future
missions. Effective explanations will better enable such
determinations.
For both challenge problem areas, it is critical to
measure explanation effectiveness. While it would be
convenient if a learned model’s explainability could be
measured automatically, an XAI system’sexplanation
effectiveness must be assessed according to how its
explanations aid human users. This requires human-
in-the-loop psychologic experiments to measure the
user’s satisfaction, mental model, task performance,
and appropriate trust. DARPA formulated an initial
explanation evaluation framework that includes po-
tential measures of explanation effectiveness (figure 5).
Exploring and refining this framework is an important
part of the XAI program’sresearchagenda.
The XAI program’s goal, concept, strategies, chal-
lenges, and evaluation framework are described in
the program’s 2016 broad agency announcement.
Figure 6 displays the XAI program’s schedule, which
consists of two phases. Phase 1 (18 months) com-
menced in May 2017 and includes initial technology
demonstrations of XAI systems. Phase 2 (30 months)
includes a sequence of evaluations against challenge
problems selected by the system developers and the
XAI evaluator. The first formal evaluations of XAI
systems took place during the fall of 2018. This article
describes the developer teams’ progress leading up to
these evaluations, whose results were presented at an XAI
program meeting during the winter of 2019.
XAI Program
Development and Progress
Figure 7 summarizes the 11 XAI Technical Area 1 (TA1)
developer teams and the TA2team[fromtheFlorida
Institute for Human and Machine Cognition (IHMC)]
that is developing the psychologic model of explanation.
Three TA1 teams are pursuing both challenge problem
areas (autonomy and data analytics), three are working on
only the former, and five are working on only the latter.
Per the strategies described in figure 2, the TA1 teams are
investigating a diverse range of techniques for developing
explainable models and explanation interfaces.
Naturalistic
Decision-Making Foundations of XAI
The objective of the IHMC team ( which includes
researchers from MacroCognition and Michigan
Technological University) is to develop and evaluate
psychologically plausible models of explanation and
develop actionable concepts, methods, measures, and
metrics for explanatory reasoning. The IHMC team is
Decision
Trees
Deep
Learning
Ensemble
Methods
Random
Forests
Learning Performance
Explainability
Learning Techniques
Statistical
Models
SVMs
AOGs
Performance vs.
Explainability
Neural Nets
Markov
Models
MLNs
HBNs
SRL
CRFs
Bayesian
Belief Nets
Graphical
Models
Figure 1. Learning Performance Versus Explainability Trade-Off for Several Categories of Learning Techniques.
46 AI MAGAZINE
Deep Learning and Security
剩余14页未读,继续阅读


















安全验证
文档复制为VIP权益,开通VIP直接复制

评论0