Beruflich Dokumente
Kultur Dokumente
Huamin Qu
Hong Kong University of Science and Technology
”
VisLab’s work on VIS for AI
iForest (IEEE VAST’18)
Non-DL EmbeddingVis (IEEE VAST’18)
DL
RuleMatrix (IEEE VAST’18) (Model Trust)
Treat as black box
3
Outline
• Motivations
5
What role is visualization playing in XAI?
6
VisLab’s work on VIS for AI
iForest (IEEE VAST’18)
Non-DL EmbeddingVis (IEEE VAST’18)
DL
RuleMatrix (IEEE VAST’18) (Model Trust)
Treat as black box
7
iForest: Interpreting Random Forests via Visual Analytics
IEEE Visual Analytics Science and Technology (VAST) 2018
a. For a data item, Decision Path Projection provides an overview of decision path similarities.
Decision Path View
b. The Feature Summary shows the summarized feature ranges for multiple selected decision paths.
Decision Path View
c. The Decision Path Flow encodes the detailed structures and feature ranges in a layer-wise manner.
Titanic Usage Scenario – Decision Path View
Positive Negative
Titanic Usage Scenario – Decision Path View
Positive Negative
Titanic Usage Scenario – Decision Path View
Positive Negative
VisLab’s work on VIS for AI
iForest (IEEE VAST’18)
Non-DL EmbeddingVis (IEEE VAST’18)
DL
RuleMatrix (IEEE VAST’18) (Model Trust)
Treat as black box
19
input
? output
21
What has the RNN learned from data?
A. map the value of a single hidden unit on data (Karpathy A. et al., 2015)
22
Our Solution: RNNVis
23
https://www.youtube.com/watch?v=0QFDNLdQ6_w
RNNVis
Our Solution
Explaining individual hidden units
Bi-graph and co-clustering
Sequence evaluation
24
RNNVis
Solution
Explaining an individual hidden unit using its most salient words
25% - 75%
9% - 91%
response
Unit: #36
Top 4 positive/negative salient words of unit 36 in
an RNN (GRU) trained on Yelp review data.
25
What Has RNN Learned from Data?
Solution
Explaining an individual hidden unit using its most salient words
mean
25% - 75%
9% - 91%
Unit #
26
What Has RNN Learned from Data?
Solution
Explaining an individual hidden unit using its most salient words
27
What Has RNN Learned from Data?
Solution
Explaining individual hidden units
Bi-graph and co-clustering
Sequence evaluation
28
What Has RNN Learned from Data?
Hidden Units
Hidden Units
Hidden Units
DL
RuleMatrix (IEEE VAST’18) (Model Trust)
Treat as black box
34
DeepTracker: Visualizing the Training Process of
Convolutional Neural Networks
ACM Transactions on Intelligent Systems and Technology (TIST), 2019
True
label: dog
Rules of thumb
loss and error rate should decrease over time; consistent increase or violent fluctuation of loss may indicate a problem
a big gap between the error rates of training and validation dataset suggest the model is over-fitting
a absence of any gap may indicate the model has a limited learning capability
the update ratio is expected to be around 1e-3 (lower suggest low lr and higher suggest high lr)
Our methods
Validation view
DGMTracker (Liu et al. VAST 2017) Seq2Seq-Vis (Strobelt et al. VAST 2018)
Other Work
GanViz ( Wang et al. IEEE PVIS 2017) DQNViz ( Wang et al. IEEE VIS 2018)
VisLab’s work on VIS for AI
iForest (IEEE VAST’18)
Non-DL EmbeddingVis (IEEE VAST’18)
DL
RuleMatrix (IEEE VAST’18) (Model Trust)
Treat as black box
44
RuleMatrix: Visualizing and Understanding
Classifiers with Rules
IEEE Visual Analytics Science and Technology (VAST), 2018
Yao Ming1, Huamin Qu1, Enrico Bertini2.
1 2
RuleMatrix
RuleMatrix
...
RuleMatrix
Model-agnostic
Induct Bayesian Rule List
RuleMatrix IF (X1 in (178.67, inf)) THEN prob: [0.0152, 0.9848]
ELSE IF (X5 in (39.376, inf)) and (X6 in (1.0258, 2.0217)) THEN prob: [0.0784, 0.9216]
The Problems of Rule List in Text Form ELSE IF (X1 in (-inf, 86.89)) THEN prob: [0.9932, 0.0068]
ELSE IF (X5 in (-inf, 23.632)) and (X7 in (-inf, 25.426)) THEN prob: [0.9850, 0.0150]
ELSE IF (X4 in (203.8, inf)) and (X7 in (-inf, 25.426)) THEN prob: [0.8426, 0.1574]
IF X1 < 4 AND 9 < X2 THEN ELSEProb = 0.8 - Features are not aligned, difficult for visual
IF (X1 in (137.52, 155.06)) and (X5 in (39.376, inf)) THEN prob: [0.0149, 0.9851]
comparison/search
ELSE IF (X1 in (155.06, 178.67)) THEN prob: [0.0675, 0.9325]
ELSE IF X3 = 1 THEN Prob =ELSE IF0.2
(X5 in (39.376, inf)) and (X7 in (36.007, inf)) THEN prob: [0.1786, 0.8214]
- Unable to view important information of
ELSE IF (X1 in (86.89, 107.17)) THEN prob: [0.9750, 0.0250]
ELSE IF X4 = 0 THEN Prob =ELSE IF0.1 each rule (supports, fidelity, etc)
(X5 in (32.169, 39.376)) and (X7 in (36.007, inf)) THEN prob: [0.1835, 0.8165]
ELSE IF (X1 in (137.52, 155.06)) and (X6 in (0.3688, 1.0258)) THEN prob: [0.2250, 0.7750]
Solution: RuleMatrix Visualization! ELSE IF (X7 in (-inf, 25.426)) THEN prob: [0.9842, 0.0158]
ELSE IF (X1 in (132.04, 137.52)) and (X6 in (0.3688, 1.0258)) THEN prob: [0.3600, 0.6400]
ELSE IF (X0 in (-inf, 4.668)) and (X5 in (23.632, 28.954)) THEN prob: [0.9884, 0.0116]
2
2
RuleMatrix
Data Flow
Negative Positive
predict by the model as Negative but wrong
Predict by the model as Positive but wrong
Case - Understand the Model
Young age AND Low BMI -> Negative
DL
RuleMatrix (IEEE VAST’18) (Model Trust)
Treat as black box
62
ATMSeer:
Increasing Transparency and Controllability in
Automated Machine Learning
Qianwen Wang, Yao Ming, Zhihua Jin, Qiaomu Shen, Dongyu Liu,
Micah J. Smith, Kalyan Veeramachaneni, Huamin Qu
63
Motivation
Make it automated
Motivation
Run long enough?
Sufficiently explore
the search space? Miss some
suitable models?
lt s ,
Prior Resu rns,
Knowledg e
e Patt hts
In s ig
Automated
Machine Learning
67
Motivation
Controllability
Automated
Machine Learning
68
ATMSeer: Increasing Transparency and Controllability
in Automated Machine Learning
Transparency
e s u l ts,
R
Prior Automatedtt e r n s,
Pa s
Knowledg s ig h t
e
Machine In
Learning
Controllability
Transparency Controllability
Analyze the searched models Automated
Modify the search space
Machine Learning 69
Designing ATMSeer
Start
Yes
Yes Yes
72
Designing ATMSeer
• Unsatisfying results
D2. Adjust
computational
• Potential to improve
budget? • Low coverage
• ……
Compu
tational
Budget
Model
Perform
ance
73
Designing ATMSeer
D3.
Reason/analyze
• Unfamiliar with the model
model choice? • Models with similar performances
Performance
Robustness
Score
Trust
74
Usage Scenarios
75
Usage Scenarios
76
Usage Scenarios
77
Other Findings
Different suitable hyperparameters for different dataset
78
Other Findings
Same dataset, different algorithms
79
VisLab’s work on VIS for AI
iForest (IEEE VAST’18)
Non-DL EmbeddingVis (IEEE VAST’18)
DL
RuleMatrix (IEEE VAST’18) (Model Trust)
Treat as black box
80
Limitations and Future Work
Limitations:
• Scalability
• Generalization
• Validation
Future Work:
• Model bias
• Explainability by analogy
• …
Thank You!
Contact:
Huamin Qu
huamin@cse.ust.hk
More Info:
http://vis.cse.ust.hk/groups/xai-vis/
82