ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1707.01943
  4. Cited By
A causal framework for explaining the predictions of black-box
  sequence-to-sequence models

A causal framework for explaining the predictions of black-box sequence-to-sequence models

6 July 2017
David Alvarez-Melis
Tommi Jaakkola
    CML
ArXivPDFHTML

Papers citing "A causal framework for explaining the predictions of black-box sequence-to-sequence models"

46 / 46 papers shown
Title
Fake News Detection After LLM Laundering: Measurement and Explanation
Fake News Detection After LLM Laundering: Measurement and Explanation
Rupak Kumar Das
Jonathan Dodge
91
0
0
29 Jan 2025
On the Probability of Necessity and Sufficiency of Explaining Graph Neural Networks: A Lower Bound Optimization Approach
On the Probability of Necessity and Sufficiency of Explaining Graph Neural Networks: A Lower Bound Optimization Approach
Ruichu Cai
Yuxuan Zhu
Xuexin Chen
Yuan Fang
Min-man Wu
Jie Qiao
Zhifeng Hao
51
7
0
31 Dec 2024
CELL your Model: Contrastive Explanations for Large Language Models
CELL your Model: Contrastive Explanations for Large Language Models
Ronny Luss
Erik Miehling
Amit Dhurandhar
47
0
0
17 Jun 2024
Explaining high-dimensional text classifiers
Explaining high-dimensional text classifiers
Odelia Melamed
Rich Caruana
23
0
0
22 Nov 2023
Towards Explainable AI Writing Assistants for Non-native English
  Speakers
Towards Explainable AI Writing Assistants for Non-native English Speakers
Yewon Kim
Mina Lee
Donghwi Kim
Sung-Ju Lee
16
4
0
05 Apr 2023
Towards Learning and Explaining Indirect Causal Effects in Neural
  Networks
Towards Learning and Explaining Indirect Causal Effects in Neural Networks
Abbaavaram Gowtham Reddy
Saketh Bachu
Harsh Nilesh Pathak
Ben Godfrey
V. Balasubramanian
V. Varshaneya
Satya Narayanan Kar
CML
31
0
0
24 Mar 2023
NxPlain: Web-based Tool for Discovery of Latent Concepts
NxPlain: Web-based Tool for Discovery of Latent Concepts
Fahim Dalvi
Nadir Durrani
Hassan Sajjad
Tamim Jaban
Musab Husaini
Ummar Abbas
15
1
0
06 Mar 2023
Understanding and Detecting Hallucinations in Neural Machine Translation
  via Model Introspection
Understanding and Detecting Hallucinations in Neural Machine Translation via Model Introspection
Weijia Xu
Sweta Agrawal
Eleftheria Briakou
Marianna J. Martindale
Marine Carpuat
HILM
27
46
0
18 Jan 2023
Influence Functions for Sequence Tagging Models
Influence Functions for Sequence Tagging Models
Sarthak Jain
Varun Manjunatha
Byron C. Wallace
A. Nenkova
TDI
35
8
0
25 Oct 2022
On the Explainability of Natural Language Processing Deep Models
On the Explainability of Natural Language Processing Deep Models
Julia El Zini
M. Awad
27
82
0
13 Oct 2022
An Interpretability Evaluation Benchmark for Pre-trained Language Models
An Interpretability Evaluation Benchmark for Pre-trained Language Models
Ya-Ming Shen
Lijie Wang
Ying Chen
Xinyan Xiao
Jing Liu
Hua-Hong Wu
37
4
0
28 Jul 2022
Leveraging Causal Inference for Explainable Automatic Program Repair
Leveraging Causal Inference for Explainable Automatic Program Repair
Jianzong Wang
Shijing Si
Z. Zhu
Xiaoyang Qu
Zhenhou Hong
Jing Xiao
27
3
0
26 May 2022
A Fine-grained Interpretability Evaluation Benchmark for Neural NLP
A Fine-grained Interpretability Evaluation Benchmark for Neural NLP
Lijie Wang
Yaozong Shen
Shu-ping Peng
Shuai Zhang
Xinyan Xiao
Hao Liu
Hongxuan Tang
Ying Chen
Hua-Hong Wu
Haifeng Wang
ELM
16
21
0
23 May 2022
Can Rationalization Improve Robustness?
Can Rationalization Improve Robustness?
Howard Chen
Jacqueline He
Karthik Narasimhan
Danqi Chen
AAML
23
40
0
25 Apr 2022
VALUE: Understanding Dialect Disparity in NLU
VALUE: Understanding Dialect Disparity in NLU
Caleb Ziems
Jiaao Chen
Camille Harris
J. Anderson
Diyi Yang
ELM
44
41
0
06 Apr 2022
Discovering Invariant Rationales for Graph Neural Networks
Discovering Invariant Rationales for Graph Neural Networks
Yingmin Wu
Xiang Wang
An Zhang
Xiangnan He
Tat-Seng Chua
OOD
AI4CE
99
224
0
30 Jan 2022
Matching Learned Causal Effects of Neural Networks with Domain Priors
Matching Learned Causal Effects of Neural Networks with Domain Priors
Sai Srinivas Kancheti
Abbavaram Gowtham Reddy
V. Balasubramanian
Amit Sharma
CML
28
12
0
24 Nov 2021
Let the CAT out of the bag: Contrastive Attributed explanations for Text
Let the CAT out of the bag: Contrastive Attributed explanations for Text
Saneem A. Chemmengath
A. Azad
Ronny Luss
Amit Dhurandhar
FAtt
34
10
0
16 Sep 2021
Counterfactual Evaluation for Explainable AI
Counterfactual Evaluation for Explainable AI
Yingqiang Ge
Shuchang Liu
Zelong Li
Shuyuan Xu
Shijie Geng
Yunqi Li
Juntao Tan
Fei Sun
Yongfeng Zhang
CML
35
13
0
05 Sep 2021
Towards Out-Of-Distribution Generalization: A Survey
Towards Out-Of-Distribution Generalization: A Survey
Jiashuo Liu
Zheyan Shen
Yue He
Xingxuan Zhang
Renzhe Xu
Han Yu
Peng Cui
CML
OOD
55
517
0
31 Aug 2021
Rationalization through Concepts
Rationalization through Concepts
Diego Antognini
Boi Faltings
FAtt
24
19
0
11 May 2021
Local Interpretations for Explainable Natural Language Processing: A
  Survey
Local Interpretations for Explainable Natural Language Processing: A Survey
Siwen Luo
Hamish Ivison
S. Han
Josiah Poon
MILM
33
48
0
20 Mar 2021
CAT-Gen: Improving Robustness in NLP Models via Controlled Adversarial
  Text Generation
CAT-Gen: Improving Robustness in NLP Models via Controlled Adversarial Text Generation
Tianlu Wang
Xuezhi Wang
Yao Qin
Ben Packer
Kang Li
Jilin Chen
Alex Beutel
Ed H. Chi
SILM
32
82
0
05 Oct 2020
Counterfactual Explanation and Causal Inference in Service of Robustness
  in Robot Control
Counterfactual Explanation and Causal Inference in Service of Robustness in Robot Control
Simón C. Smith
S. Ramamoorthy
23
13
0
18 Sep 2020
Interactive Visual Study of Multiple Attributes Learning Model of X-Ray
  Scattering Images
Interactive Visual Study of Multiple Attributes Learning Model of X-Ray Scattering Images
Xinyi Huang
Suphanut Jamonnak
Ye Zhao
Boyu Wang
Minh Hoai
Kevin Yager
Wei-ping Xu
30
5
0
03 Sep 2020
Survey of XAI in digital pathology
Survey of XAI in digital pathology
Milda Pocevičiūtė
Gabriel Eilertsen
Claes Lundström
8
56
0
14 Aug 2020
BERTology Meets Biology: Interpreting Attention in Protein Language
  Models
BERTology Meets Biology: Interpreting Attention in Protein Language Models
Jesse Vig
Ali Madani
L. Varshney
Caiming Xiong
R. Socher
Nazneen Rajani
29
288
0
26 Jun 2020
Generative causal explanations of black-box classifiers
Generative causal explanations of black-box classifiers
Matthew R. O’Shaughnessy
Gregory H. Canal
Marissa Connor
Mark A. Davenport
Christopher Rozell
CML
30
73
0
24 Jun 2020
Causal Interpretability for Machine Learning -- Problems, Methods and
  Evaluation
Causal Interpretability for Machine Learning -- Problems, Methods and Evaluation
Raha Moraffah
Mansooreh Karami
Ruocheng Guo
A. Raglin
Huan Liu
CML
ELM
XAI
27
213
0
09 Mar 2020
Neural Machine Translation: A Review and Survey
Neural Machine Translation: A Review and Survey
Felix Stahlberg
3DV
AI4TS
MedIm
20
311
0
04 Dec 2019
Weight of Evidence as a Basis for Human-Oriented Explanations
Weight of Evidence as a Basis for Human-Oriented Explanations
David Alvarez-Melis
Hal Daumé
Jennifer Wortman Vaughan
Hanna M. Wallach
XAI
FAtt
21
20
0
29 Oct 2019
A Game Theoretic Approach to Class-wise Selective Rationalization
A Game Theoretic Approach to Class-wise Selective Rationalization
Shiyu Chang
Yang Zhang
Mo Yu
Tommi Jaakkola
22
60
0
28 Oct 2019
MonoNet: Towards Interpretable Models by Learning Monotonic Features
MonoNet: Towards Interpretable Models by Learning Monotonic Features
An-phi Nguyen
María Rodríguez Martínez
FAtt
16
13
0
30 Sep 2019
On Model Stability as a Function of Random Seed
On Model Stability as a Function of Random Seed
Pranava Madhyastha
Dhruv Batra
42
61
0
23 Sep 2019
Evaluating Gender Bias in Machine Translation
Evaluating Gender Bias in Machine Translation
Gabriel Stanovsky
Noah A. Smith
Luke Zettlemoyer
13
393
0
03 Jun 2019
Interpretable Neural Predictions with Differentiable Binary Variables
Interpretable Neural Predictions with Differentiable Binary Variables
Jasmijn Bastings
Wilker Aziz
Ivan Titov
23
211
0
20 May 2019
Semantics Preserving Adversarial Learning
Semantics Preserving Adversarial Learning
Ousmane Amadou Dia
Elnaz Barshan
Reza Babanezhad
AAML
GAN
24
2
0
10 Mar 2019
Attention is not Explanation
Attention is not Explanation
Sarthak Jain
Byron C. Wallace
FAtt
31
1,298
0
26 Feb 2019
e-SNLI: Natural Language Inference with Natural Language Explanations
e-SNLI: Natural Language Inference with Natural Language Explanations
Oana-Maria Camburu
Tim Rocktaschel
Thomas Lukasiewicz
Phil Blunsom
LRM
257
620
0
04 Dec 2018
An Operation Sequence Model for Explainable Neural Machine Translation
An Operation Sequence Model for Explainable Neural Machine Translation
Felix Stahlberg
Danielle Saunders
Bill Byrne
LRM
MILM
40
29
0
29 Aug 2018
On the Robustness of Interpretability Methods
On the Robustness of Interpretability Methods
David Alvarez-Melis
Tommi Jaakkola
30
521
0
21 Jun 2018
Towards Robust Interpretability with Self-Explaining Neural Networks
Towards Robust Interpretability with Self-Explaining Neural Networks
David Alvarez-Melis
Tommi Jaakkola
MILM
XAI
29
932
0
20 Jun 2018
AGI Safety Literature Review
AGI Safety Literature Review
Tom Everitt
G. Lea
Marcus Hutter
AI4CE
34
115
0
03 May 2018
Seq2Seq-Vis: A Visual Debugging Tool for Sequence-to-Sequence Models
Seq2Seq-Vis: A Visual Debugging Tool for Sequence-to-Sequence Models
Hendrik Strobelt
Sebastian Gehrmann
M. Behrisch
Adam Perer
Hanspeter Pfister
Alexander M. Rush
VLM
HAI
31
239
0
25 Apr 2018
Generating Natural Adversarial Examples
Generating Natural Adversarial Examples
Zhengli Zhao
Dheeru Dua
Sameer Singh
GAN
AAML
38
596
0
31 Oct 2017
OpenNMT: Open-Source Toolkit for Neural Machine Translation
OpenNMT: Open-Source Toolkit for Neural Machine Translation
Guillaume Klein
Yoon Kim
Yuntian Deng
Jean Senellart
Alexander M. Rush
259
1,896
0
10 Jan 2017
1