Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1707.01943
Cited By
A causal framework for explaining the predictions of black-box sequence-to-sequence models
6 July 2017
David Alvarez-Melis
Tommi Jaakkola
CML
Re-assign community
ArXiv
PDF
HTML
Papers citing
"A causal framework for explaining the predictions of black-box sequence-to-sequence models"
46 / 46 papers shown
Title
Fake News Detection After LLM Laundering: Measurement and Explanation
Rupak Kumar Das
Jonathan Dodge
91
0
0
29 Jan 2025
On the Probability of Necessity and Sufficiency of Explaining Graph Neural Networks: A Lower Bound Optimization Approach
Ruichu Cai
Yuxuan Zhu
Xuexin Chen
Yuan Fang
Min-man Wu
Jie Qiao
Zhifeng Hao
51
7
0
31 Dec 2024
CELL your Model: Contrastive Explanations for Large Language Models
Ronny Luss
Erik Miehling
Amit Dhurandhar
47
0
0
17 Jun 2024
Explaining high-dimensional text classifiers
Odelia Melamed
Rich Caruana
23
0
0
22 Nov 2023
Towards Explainable AI Writing Assistants for Non-native English Speakers
Yewon Kim
Mina Lee
Donghwi Kim
Sung-Ju Lee
16
4
0
05 Apr 2023
Towards Learning and Explaining Indirect Causal Effects in Neural Networks
Abbaavaram Gowtham Reddy
Saketh Bachu
Harsh Nilesh Pathak
Ben Godfrey
V. Balasubramanian
V. Varshaneya
Satya Narayanan Kar
CML
31
0
0
24 Mar 2023
NxPlain: Web-based Tool for Discovery of Latent Concepts
Fahim Dalvi
Nadir Durrani
Hassan Sajjad
Tamim Jaban
Musab Husaini
Ummar Abbas
15
1
0
06 Mar 2023
Understanding and Detecting Hallucinations in Neural Machine Translation via Model Introspection
Weijia Xu
Sweta Agrawal
Eleftheria Briakou
Marianna J. Martindale
Marine Carpuat
HILM
27
46
0
18 Jan 2023
Influence Functions for Sequence Tagging Models
Sarthak Jain
Varun Manjunatha
Byron C. Wallace
A. Nenkova
TDI
35
8
0
25 Oct 2022
On the Explainability of Natural Language Processing Deep Models
Julia El Zini
M. Awad
27
82
0
13 Oct 2022
An Interpretability Evaluation Benchmark for Pre-trained Language Models
Ya-Ming Shen
Lijie Wang
Ying Chen
Xinyan Xiao
Jing Liu
Hua-Hong Wu
37
4
0
28 Jul 2022
Leveraging Causal Inference for Explainable Automatic Program Repair
Jianzong Wang
Shijing Si
Z. Zhu
Xiaoyang Qu
Zhenhou Hong
Jing Xiao
27
3
0
26 May 2022
A Fine-grained Interpretability Evaluation Benchmark for Neural NLP
Lijie Wang
Yaozong Shen
Shu-ping Peng
Shuai Zhang
Xinyan Xiao
Hao Liu
Hongxuan Tang
Ying Chen
Hua-Hong Wu
Haifeng Wang
ELM
16
21
0
23 May 2022
Can Rationalization Improve Robustness?
Howard Chen
Jacqueline He
Karthik Narasimhan
Danqi Chen
AAML
23
40
0
25 Apr 2022
VALUE: Understanding Dialect Disparity in NLU
Caleb Ziems
Jiaao Chen
Camille Harris
J. Anderson
Diyi Yang
ELM
44
41
0
06 Apr 2022
Discovering Invariant Rationales for Graph Neural Networks
Yingmin Wu
Xiang Wang
An Zhang
Xiangnan He
Tat-Seng Chua
OOD
AI4CE
99
224
0
30 Jan 2022
Matching Learned Causal Effects of Neural Networks with Domain Priors
Sai Srinivas Kancheti
Abbavaram Gowtham Reddy
V. Balasubramanian
Amit Sharma
CML
28
12
0
24 Nov 2021
Let the CAT out of the bag: Contrastive Attributed explanations for Text
Saneem A. Chemmengath
A. Azad
Ronny Luss
Amit Dhurandhar
FAtt
34
10
0
16 Sep 2021
Counterfactual Evaluation for Explainable AI
Yingqiang Ge
Shuchang Liu
Zelong Li
Shuyuan Xu
Shijie Geng
Yunqi Li
Juntao Tan
Fei Sun
Yongfeng Zhang
CML
35
13
0
05 Sep 2021
Towards Out-Of-Distribution Generalization: A Survey
Jiashuo Liu
Zheyan Shen
Yue He
Xingxuan Zhang
Renzhe Xu
Han Yu
Peng Cui
CML
OOD
55
517
0
31 Aug 2021
Rationalization through Concepts
Diego Antognini
Boi Faltings
FAtt
24
19
0
11 May 2021
Local Interpretations for Explainable Natural Language Processing: A Survey
Siwen Luo
Hamish Ivison
S. Han
Josiah Poon
MILM
33
48
0
20 Mar 2021
CAT-Gen: Improving Robustness in NLP Models via Controlled Adversarial Text Generation
Tianlu Wang
Xuezhi Wang
Yao Qin
Ben Packer
Kang Li
Jilin Chen
Alex Beutel
Ed H. Chi
SILM
32
82
0
05 Oct 2020
Counterfactual Explanation and Causal Inference in Service of Robustness in Robot Control
Simón C. Smith
S. Ramamoorthy
23
13
0
18 Sep 2020
Interactive Visual Study of Multiple Attributes Learning Model of X-Ray Scattering Images
Xinyi Huang
Suphanut Jamonnak
Ye Zhao
Boyu Wang
Minh Hoai
Kevin Yager
Wei-ping Xu
30
5
0
03 Sep 2020
Survey of XAI in digital pathology
Milda Pocevičiūtė
Gabriel Eilertsen
Claes Lundström
8
56
0
14 Aug 2020
BERTology Meets Biology: Interpreting Attention in Protein Language Models
Jesse Vig
Ali Madani
L. Varshney
Caiming Xiong
R. Socher
Nazneen Rajani
29
288
0
26 Jun 2020
Generative causal explanations of black-box classifiers
Matthew R. O’Shaughnessy
Gregory H. Canal
Marissa Connor
Mark A. Davenport
Christopher Rozell
CML
30
73
0
24 Jun 2020
Causal Interpretability for Machine Learning -- Problems, Methods and Evaluation
Raha Moraffah
Mansooreh Karami
Ruocheng Guo
A. Raglin
Huan Liu
CML
ELM
XAI
27
213
0
09 Mar 2020
Neural Machine Translation: A Review and Survey
Felix Stahlberg
3DV
AI4TS
MedIm
20
311
0
04 Dec 2019
Weight of Evidence as a Basis for Human-Oriented Explanations
David Alvarez-Melis
Hal Daumé
Jennifer Wortman Vaughan
Hanna M. Wallach
XAI
FAtt
21
20
0
29 Oct 2019
A Game Theoretic Approach to Class-wise Selective Rationalization
Shiyu Chang
Yang Zhang
Mo Yu
Tommi Jaakkola
22
60
0
28 Oct 2019
MonoNet: Towards Interpretable Models by Learning Monotonic Features
An-phi Nguyen
María Rodríguez Martínez
FAtt
16
13
0
30 Sep 2019
On Model Stability as a Function of Random Seed
Pranava Madhyastha
Dhruv Batra
42
61
0
23 Sep 2019
Evaluating Gender Bias in Machine Translation
Gabriel Stanovsky
Noah A. Smith
Luke Zettlemoyer
13
393
0
03 Jun 2019
Interpretable Neural Predictions with Differentiable Binary Variables
Jasmijn Bastings
Wilker Aziz
Ivan Titov
23
211
0
20 May 2019
Semantics Preserving Adversarial Learning
Ousmane Amadou Dia
Elnaz Barshan
Reza Babanezhad
AAML
GAN
24
2
0
10 Mar 2019
Attention is not Explanation
Sarthak Jain
Byron C. Wallace
FAtt
31
1,298
0
26 Feb 2019
e-SNLI: Natural Language Inference with Natural Language Explanations
Oana-Maria Camburu
Tim Rocktaschel
Thomas Lukasiewicz
Phil Blunsom
LRM
257
620
0
04 Dec 2018
An Operation Sequence Model for Explainable Neural Machine Translation
Felix Stahlberg
Danielle Saunders
Bill Byrne
LRM
MILM
40
29
0
29 Aug 2018
On the Robustness of Interpretability Methods
David Alvarez-Melis
Tommi Jaakkola
30
521
0
21 Jun 2018
Towards Robust Interpretability with Self-Explaining Neural Networks
David Alvarez-Melis
Tommi Jaakkola
MILM
XAI
29
932
0
20 Jun 2018
AGI Safety Literature Review
Tom Everitt
G. Lea
Marcus Hutter
AI4CE
34
115
0
03 May 2018
Seq2Seq-Vis: A Visual Debugging Tool for Sequence-to-Sequence Models
Hendrik Strobelt
Sebastian Gehrmann
M. Behrisch
Adam Perer
Hanspeter Pfister
Alexander M. Rush
VLM
HAI
31
239
0
25 Apr 2018
Generating Natural Adversarial Examples
Zhengli Zhao
Dheeru Dua
Sameer Singh
GAN
AAML
38
596
0
31 Oct 2017
OpenNMT: Open-Source Toolkit for Neural Machine Translation
Guillaume Klein
Yoon Kim
Yuntian Deng
Jean Senellart
Alexander M. Rush
259
1,896
0
10 Jan 2017
1