Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2010.05419
Cited By
Gradient-based Analysis of NLP Models is Manipulable
12 October 2020
Junlin Wang
Jens Tuyls
Eric Wallace
Sameer Singh
AAML
FAtt
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Gradient-based Analysis of NLP Models is Manipulable"
16 / 16 papers shown
Title
Gradient based Feature Attribution in Explainable AI: A Technical Review
Yongjie Wang
Tong Zhang
Xu Guo
Zhiqi Shen
XAI
29
19
0
15 Mar 2024
Neighboring Words Affect Human Interpretation of Saliency Explanations
Tim Dockhorn
Yaoliang Yu
Heike Adel
Mahdi Zolnouri
V. Nia
FAtt
MILM
41
3
0
04 May 2023
Improving Prediction Performance and Model Interpretability through Attention Mechanisms from Basic and Applied Research Perspectives
Shunsuke Kitada
FaML
HAI
AI4CE
32
0
0
24 Mar 2023
Tell Model Where to Attend: Improving Interpretability of Aspect-Based Sentiment Classification via Small Explanation Annotations
Zhenxiao Cheng
Jie Zhou
Wen Wu
Qin Chen
Liang He
32
3
0
21 Feb 2023
Identifying the Source of Vulnerability in Explanation Discrepancy: A Case Study in Neural Text Classification
Ruixuan Tang
Hanjie Chen
Yangfeng Ji
AAML
FAtt
32
2
0
10 Dec 2022
Explainable Artificial Intelligence Applications in Cyber Security: State-of-the-Art in Research
Zhibo Zhang
H. A. Hamadi
Ernesto Damiani
C. Yeun
Fatma Taher
AAML
34
148
0
31 Aug 2022
SBERT studies Meaning Representations: Decomposing Sentence Embeddings into Explainable Semantic Features
Juri Opitz
Anette Frank
34
33
0
14 Jun 2022
Towards a Theory of Faithfulness: Faithful Explanations of Differentiable Classifiers over Continuous Data
Nico Potyka
Xiang Yin
Francesca Toni
FAtt
22
2
0
19 May 2022
It Takes Two Flints to Make a Fire: Multitask Learning of Neural Relation and Explanation Classifiers
Zheng Tang
Mihai Surdeanu
27
6
0
25 Apr 2022
Interpretation of Black Box NLP Models: A Survey
Shivani Choudhary
N. Chatterjee
S. K. Saha
FAtt
34
10
0
31 Mar 2022
Framework for Evaluating Faithfulness of Local Explanations
S. Dasgupta
Nave Frost
Michal Moshkovitz
FAtt
119
61
0
01 Feb 2022
Human Interpretation of Saliency-based Explanation Over Text
Hendrik Schuff
Alon Jacovi
Heike Adel
Yoav Goldberg
Ngoc Thang Vu
MILM
XAI
FAtt
148
39
0
27 Jan 2022
How Reliable are Model Diagnostics?
V. Aribandi
Yi Tay
Donald Metzler
19
19
0
12 May 2021
Towards Robust Explanations for Deep Neural Networks
Ann-Kathrin Dombrowski
Christopher J. Anders
K. Müller
Pan Kessel
FAtt
35
63
0
18 Dec 2020
Reliable Post hoc Explanations: Modeling Uncertainty in Explainability
Dylan Slack
Sophie Hilgard
Sameer Singh
Himabindu Lakkaraju
FAtt
29
162
0
11 Aug 2020
Towards A Rigorous Science of Interpretable Machine Learning
Finale Doshi-Velez
Been Kim
XAI
FaML
257
3,696
0
28 Feb 2017
1