ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2010.05419
  4. Cited By
Gradient-based Analysis of NLP Models is Manipulable

Gradient-based Analysis of NLP Models is Manipulable

12 October 2020
Junlin Wang
Jens Tuyls
Eric Wallace
Sameer Singh
    AAML
    FAtt
ArXivPDFHTML

Papers citing "Gradient-based Analysis of NLP Models is Manipulable"

16 / 16 papers shown
Title
Gradient based Feature Attribution in Explainable AI: A Technical Review
Gradient based Feature Attribution in Explainable AI: A Technical Review
Yongjie Wang
Tong Zhang
Xu Guo
Zhiqi Shen
XAI
29
19
0
15 Mar 2024
Neighboring Words Affect Human Interpretation of Saliency Explanations
Neighboring Words Affect Human Interpretation of Saliency Explanations
Tim Dockhorn
Yaoliang Yu
Heike Adel
Mahdi Zolnouri
V. Nia
FAtt
MILM
41
3
0
04 May 2023
Improving Prediction Performance and Model Interpretability through
  Attention Mechanisms from Basic and Applied Research Perspectives
Improving Prediction Performance and Model Interpretability through Attention Mechanisms from Basic and Applied Research Perspectives
Shunsuke Kitada
FaML
HAI
AI4CE
32
0
0
24 Mar 2023
Tell Model Where to Attend: Improving Interpretability of Aspect-Based
  Sentiment Classification via Small Explanation Annotations
Tell Model Where to Attend: Improving Interpretability of Aspect-Based Sentiment Classification via Small Explanation Annotations
Zhenxiao Cheng
Jie Zhou
Wen Wu
Qin Chen
Liang He
32
3
0
21 Feb 2023
Identifying the Source of Vulnerability in Explanation Discrepancy: A
  Case Study in Neural Text Classification
Identifying the Source of Vulnerability in Explanation Discrepancy: A Case Study in Neural Text Classification
Ruixuan Tang
Hanjie Chen
Yangfeng Ji
AAML
FAtt
32
2
0
10 Dec 2022
Explainable Artificial Intelligence Applications in Cyber Security:
  State-of-the-Art in Research
Explainable Artificial Intelligence Applications in Cyber Security: State-of-the-Art in Research
Zhibo Zhang
H. A. Hamadi
Ernesto Damiani
C. Yeun
Fatma Taher
AAML
34
148
0
31 Aug 2022
SBERT studies Meaning Representations: Decomposing Sentence Embeddings
  into Explainable Semantic Features
SBERT studies Meaning Representations: Decomposing Sentence Embeddings into Explainable Semantic Features
Juri Opitz
Anette Frank
34
33
0
14 Jun 2022
Towards a Theory of Faithfulness: Faithful Explanations of
  Differentiable Classifiers over Continuous Data
Towards a Theory of Faithfulness: Faithful Explanations of Differentiable Classifiers over Continuous Data
Nico Potyka
Xiang Yin
Francesca Toni
FAtt
22
2
0
19 May 2022
It Takes Two Flints to Make a Fire: Multitask Learning of Neural
  Relation and Explanation Classifiers
It Takes Two Flints to Make a Fire: Multitask Learning of Neural Relation and Explanation Classifiers
Zheng Tang
Mihai Surdeanu
27
6
0
25 Apr 2022
Interpretation of Black Box NLP Models: A Survey
Interpretation of Black Box NLP Models: A Survey
Shivani Choudhary
N. Chatterjee
S. K. Saha
FAtt
34
10
0
31 Mar 2022
Framework for Evaluating Faithfulness of Local Explanations
Framework for Evaluating Faithfulness of Local Explanations
S. Dasgupta
Nave Frost
Michal Moshkovitz
FAtt
119
61
0
01 Feb 2022
Human Interpretation of Saliency-based Explanation Over Text
Human Interpretation of Saliency-based Explanation Over Text
Hendrik Schuff
Alon Jacovi
Heike Adel
Yoav Goldberg
Ngoc Thang Vu
MILM
XAI
FAtt
148
39
0
27 Jan 2022
How Reliable are Model Diagnostics?
How Reliable are Model Diagnostics?
V. Aribandi
Yi Tay
Donald Metzler
19
19
0
12 May 2021
Towards Robust Explanations for Deep Neural Networks
Towards Robust Explanations for Deep Neural Networks
Ann-Kathrin Dombrowski
Christopher J. Anders
K. Müller
Pan Kessel
FAtt
35
63
0
18 Dec 2020
Reliable Post hoc Explanations: Modeling Uncertainty in Explainability
Reliable Post hoc Explanations: Modeling Uncertainty in Explainability
Dylan Slack
Sophie Hilgard
Sameer Singh
Himabindu Lakkaraju
FAtt
29
162
0
11 Aug 2020
Towards A Rigorous Science of Interpretable Machine Learning
Towards A Rigorous Science of Interpretable Machine Learning
Finale Doshi-Velez
Been Kim
XAI
FaML
257
3,696
0
28 Feb 2017
1