Explain, Edit, and Understand:
Rethinking User Study Design
for Evaluating Model Explanations

Siddhant Arora
Danish Pruthi
Norman Sadeh
William W. Cohen
CMU
CMU
CMU
Google AI
Zachary C. Lipton
Graham Neubig
CMU
CMU
Accepted at AAAI, 2022

[Paper]
[Demo]
[Code]



In attempts to "explain" predictions of machine learning models, researchers have proposed hundreds of techniques for attributing predictions to features that are deemed important. While these attributions are often claimed to hold the potential to improve human "understanding" of the models, surprisingly little work explicitly evaluates progress towards this aspiration. In this paper, we conduct a crowdsourcing study, where participants interact with deception detection models that have been trained to distinguish between genuine and fake hotel reviews. They are challenged both to simulate the model on fresh reviews, and to edit reviews with the goal of lowering the probability of the originally predicted class. Successful manipulations would lead to an adversarial example. During the training (but not the test) phase, input spans are highlighted to communicate salience. Through our evaluation, we observe that for a linear bag-of-words model, participants with access to the feature coefficients during training are able to cause a larger reduction in model confidence in the testing phase when compared to the no-explanation control. For the BERT-based classifier, popular local explanations do not improve their ability to reduce the model confidence over the no-explanation case. Remarkably, when the explanation for the BERT model is given by the (global) attributions of a linear model trained to imitate the BERT model, people can effectively manipulate the model.


Demo Video



Paper and Bibtex

[Paper]

Citation
 
Arora, S. and Pruthi, D. and Sadeh, N. and Cohen, W.W. and Lipton, Z.C. and Neubig, G. 2022.
Explain, Edit, and Understand: Rethinking User Study Design for Evaluating Model Explanations.
In AAAI Conference on Artificial Intelligence (AAAI-22).

[Bibtex]
@inproceedings{arora2022explain,
  title={Explain, Edit, and Understand: Rethinking User Study Design for Evaluating Model Explanations},
  author={Arora, Siddhant and Pruthi, Danish and Sadeh, Norman and Cohen, William W. and Lipton, Zachary C. and Neubig, Graham},
  booktitle={In 36th AAAI Conference on Artificial Intelligence},
  year={2022}}