Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2004.03685
Cited By
Towards Faithfully Interpretable NLP Systems: How should we define and evaluate faithfulness?
7 April 2020
Alon Jacovi
Yoav Goldberg
XAI
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Towards Faithfully Interpretable NLP Systems: How should we define and evaluate faithfulness?"
50 / 381 papers shown
Title
EiX-GNN : Concept-level eigencentrality explainer for graph neural networks
Adrien Raison
Pascal Bourdon
David Helbert
17
1
0
07 Jun 2022
Order-sensitive Shapley Values for Evaluating Conceptual Soundness of NLP Models
Kaiji Lu
Anupam Datta
23
0
0
01 Jun 2022
Attribution-based Explanations that Provide Recourse Cannot be Robust
H. Fokkema
R. D. Heide
T. Erven
FAtt
47
18
0
31 May 2022
Comparing interpretation methods in mental state decoding analyses with deep learning models
A. Thomas
Christopher Ré
R. Poldrack
AI4CE
39
2
0
31 May 2022
CEBaB: Estimating the Causal Effects of Real-World Concepts on NLP Model Behavior
Eldar David Abraham
Karel DÓosterlinck
Amir Feder
Y. Gat
Atticus Geiger
Christopher Potts
Roi Reichart
Zhengxuan Wu
CML
36
44
0
27 May 2022
Investigating the Benefits of Free-Form Rationales
Jiao Sun
Swabha Swayamdipta
Jonathan May
Xuezhe Ma
24
14
0
25 May 2022
Logical Satisfiability of Counterfactuals for Faithful Explanations in NLI
Suzanna Sia
Anton Belyy
Amjad Almahairi
Madian Khabsa
Luke Zettlemoyer
Lambert Mathias
LRM
27
13
0
25 May 2022
Interpretation Quality Score for Measuring the Quality of interpretability methods
Sean Xie
Soroush Vosoughi
Saeed Hassanpour
XAI
19
5
0
24 May 2022
A Fine-grained Interpretability Evaluation Benchmark for Neural NLP
Lijie Wang
Yaozong Shen
Shu-ping Peng
Shuai Zhang
Xinyan Xiao
Hao Liu
Hongxuan Tang
Ying-Cong Chen
Hua Wu
Haifeng Wang
ELM
19
21
0
23 May 2022
Argumentative Explanations for Pattern-Based Text Classifiers
Piyawat Lertvittayakumjorn
Francesca Toni
45
4
0
22 May 2022
A Song of (Dis)agreement: Evaluating the Evaluation of Explainable Artificial Intelligence in Natural Language Processing
Michael Neely
Stefan F. Schouten
Maurits J. R. Bleeker
Ana Lucic
XAI
27
16
0
09 May 2022
ExSum: From Local Explanations to Model Understanding
Yilun Zhou
Marco Tulio Ribeiro
J. Shah
FAtt
LRM
27
25
0
30 Apr 2022
Can Rationalization Improve Robustness?
Howard Chen
Jacqueline He
Karthik Narasimhan
Danqi Chen
AAML
31
40
0
25 Apr 2022
Locally Aggregated Feature Attribution on Natural Language Model Understanding
Shenmin Zhang
Jin Wang
Haitao Jiang
Rui Song
FAtt
24
3
0
22 Apr 2022
Learning to Scaffold: Optimizing Model Explanations for Teaching
Patrick Fernandes
Marcos Vinícius Treviso
Danish Pruthi
André F. T. Martins
Graham Neubig
FAtt
30
22
0
22 Apr 2022
Pathologies of Pre-trained Language Models in Few-shot Fine-tuning
Hanjie Chen
Guoqing Zheng
Ahmed Hassan Awadallah
Yangfeng Ji
AI4MH
23
3
0
17 Apr 2022
Calibrating Trust of Multi-Hop Question Answering Systems with Decompositional Probes
Kaige Xie
Sarah Wiegreffe
Mark O. Riedl
ReLM
24
12
0
16 Apr 2022
A Comparative Study of Faithfulness Metrics for Model Interpretability Methods
Chun Sik Chan
Huanqi Kong
Guanqing Liang
19
50
0
12 Apr 2022
A Multilingual Perspective Towards the Evaluation of Attribution Methods in Natural Language Inference
Kerem Zaman
Yonatan Belinkov
29
8
0
11 Apr 2022
ProtoTEx: Explaining Model Decisions with Prototype Tensors
Anubrata Das
Chitrank Gupta
Venelin Kovatchev
Matthew Lease
Junjie Li
34
27
0
11 Apr 2022
Using Interactive Feedback to Improve the Accuracy and Explainability of Question Answering Systems Post-Deployment
Zichao Li
Prakhar Sharma
Xing Han Lù
Jackie C.K. Cheung
Siva Reddy
HAI
25
26
0
06 Apr 2022
Interpretation of Black Box NLP Models: A Survey
Shivani Choudhary
N. Chatterjee
S. K. Saha
FAtt
34
10
0
31 Mar 2022
STaR: Bootstrapping Reasoning With Reasoning
E. Zelikman
Yuhuai Wu
Jesse Mu
Noah D. Goodman
ReLM
LRM
44
432
0
28 Mar 2022
A Unified Study of Machine Learning Explanation Evaluation Metrics
Yipei Wang
Xiaoqian Wang
XAI
19
7
0
27 Mar 2022
Effective Explanations for Entity Resolution Models
Tommaso Teofili
D. Firmani
Nick Koudas
Vincenzo Martello
P. Merialdo
D. Srivastava
AAML
10
16
0
24 Mar 2022
Teaching language models to support answers with verified quotes
Jacob Menick
Maja Trebacz
Vladimir Mikulik
John Aslanides
Francis Song
...
Mia Glaese
Susannah Young
Lucy Campbell-Gillingham
G. Irving
Nat McAleese
ELM
RALM
251
259
0
21 Mar 2022
Towards Explainable Evaluation Metrics for Natural Language Generation
Christoph Leiter
Piyawat Lertvittayakumjorn
M. Fomicheva
Wei-Ye Zhao
Yang Gao
Steffen Eger
AAML
ELM
30
20
0
21 Mar 2022
FaiRR: Faithful and Robust Deductive Reasoning over Natural Language
Soumya Sanyal
Harman Singh
Xiang Ren
ReLM
LRM
32
45
0
19 Mar 2022
Explainability in Graph Neural Networks: An Experimental Survey
Peibo Li
Yixing Yang
Maurice Pagnucco
Yang Song
29
31
0
17 Mar 2022
A Novel Perspective to Look At Attention: Bi-level Attention-based Explainable Topic Modeling for News Classification
Dairui Liu
Derek Greene
Ruihai Dong
28
10
0
14 Mar 2022
Evaluating Explainable AI on a Multi-Modal Medical Imaging Task: Can Existing Algorithms Fulfill Clinical Requirements?
Weina Jin
Xiaoxiao Li
Ghassan Hamarneh
27
51
0
12 Mar 2022
Neuro-symbolic Natural Logic with Introspective Revision for Natural Language Inference
Yufei Feng
Xiaoyu Yang
Xiao-Dan Zhu
Michael A. Greenspan
LRM
NAI
50
11
0
09 Mar 2022
An Empirical Study on Explanations in Out-of-Domain Settings
G. Chrysostomou
Nikolaos Aletras
LRM
17
27
0
28 Feb 2022
Hierarchical Interpretation of Neural Text Classification
Hanqi Yan
Lin Gui
Yulan He
45
14
0
20 Feb 2022
Guidelines and Evaluation of Clinical Explainable AI in Medical Image Analysis
Weina Jin
Xiaoxiao Li
M. Fatehi
Ghassan Hamarneh
ELM
XAI
42
87
0
16 Feb 2022
Don't Lie to Me! Robust and Efficient Explainability with Verified Perturbation Analysis
Thomas Fel
Mélanie Ducoffe
David Vigouroux
Rémi Cadène
Mikael Capelle
C. Nicodeme
Thomas Serre
AAML
26
41
0
15 Feb 2022
DermX: an end-to-end framework for explainable automated dermatological diagnosis
Raluca Jalaboi
F. Faye
Mauricio Orbes-Arteaga
D. Jørgensen
Ole Winther
A. Galimzianova
MedIm
19
17
0
14 Feb 2022
InterpretTime: a new approach for the systematic evaluation of neural-network interpretability in time series classification
Hugues Turbé
Mina Bjelogrlic
Christian Lovis
G. Mengaldo
AI4TS
27
6
0
11 Feb 2022
Exploring Transformer Backbones for Heterogeneous Treatment Effect Estimation
Yi-Fan Zhang
Hanlin Zhang
Zachary Chase Lipton
Li Erran Li
Eric P. Xing
OODD
28
30
0
02 Feb 2022
Rethinking Attention-Model Explainability through Faithfulness Violation Test
Yue Liu
Haoliang Li
Yangyang Guo
Chen Kong
Jing Li
Shiqi Wang
FAtt
121
43
0
28 Jan 2022
Human Interpretation of Saliency-based Explanation Over Text
Hendrik Schuff
Alon Jacovi
Heike Adel
Yoav Goldberg
Ngoc Thang Vu
MILM
XAI
FAtt
148
39
0
27 Jan 2022
Diagnosing AI Explanation Methods with Folk Concepts of Behavior
Alon Jacovi
Jasmijn Bastings
Sebastian Gehrmann
Yoav Goldberg
Katja Filippova
36
15
0
27 Jan 2022
From Anecdotal Evidence to Quantitative Evaluation Methods: A Systematic Review on Evaluating Explainable AI
Meike Nauta
Jan Trienes
Shreyasi Pathak
Elisa Nguyen
Michelle Peters
Yasmin Schmitt
Jorg Schlotterer
M. V. Keulen
C. Seifert
ELM
XAI
28
398
0
20 Jan 2022
Natural Language Deduction through Search over Statement Compositions
Kaj Bostrom
Zayne Sprague
Swarat Chaudhuri
Greg Durrett
ReLM
LRM
27
46
0
16 Jan 2022
On Sensitivity of Deep Learning Based Text Classification Algorithms to Practical Input Perturbations
Aamir Miyajiwala
Arnav Ladkat
Samiksha Jagadale
Raviraj Joshi
AAML
17
7
0
02 Jan 2022
UNIREX: A Unified Learning Framework for Language Model Rationale Extraction
Aaron Chan
Maziar Sanjabi
Lambert Mathias
L Tan
Shaoliang Nie
Xiaochang Peng
Xiang Ren
Hamed Firooz
43
42
0
16 Dec 2021
Sparse Interventions in Language Models with Differentiable Masking
Nicola De Cao
Leon Schmid
Dieuwke Hupkes
Ivan Titov
40
27
0
13 Dec 2021
What I Cannot Predict, I Do Not Understand: A Human-Centered Evaluation Framework for Explainability Methods
Julien Colin
Thomas Fel
Rémi Cadène
Thomas Serre
33
101
0
06 Dec 2021
Explainable Deep Learning in Healthcare: A Methodological Survey from an Attribution View
Di Jin
Elena Sergeeva
W. Weng
Geeticka Chauhan
Peter Szolovits
OOD
41
55
0
05 Dec 2021
Inducing Causal Structure for Interpretable Neural Networks
Atticus Geiger
Zhengxuan Wu
Hanson Lu
J. Rozner
Elisa Kreiss
Thomas Icard
Noah D. Goodman
Christopher Potts
CML
OOD
35
71
0
01 Dec 2021
Previous
1
2
3
4
5
6
7
8
Next