Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2102.12781
Cited By
Do Input Gradients Highlight Discriminative Features?
25 February 2021
Harshay Shah
Prateek Jain
Praneeth Netrapalli
AAML
FAtt
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Do Input Gradients Highlight Discriminative Features?"
42 / 42 papers shown
Title
Axiomatic Explainer Globalness via Optimal Transport
Davin Hill
Josh Bone
A. Masoomi
Max Torop
Jennifer Dy
97
1
0
13 Mar 2025
Concept Bottleneck Language Models For protein design
Aya Abdelsalam Ismail
Tuomas Oikarinen
Amy Wang
Julius Adebayo
Samuel Stanton
...
J. Kleinhenz
Allen Goodman
H. C. Bravo
Kyunghyun Cho
Nathan C. Frey
32
4
0
09 Nov 2024
Unlearning-based Neural Interpretations
Ching Lam Choi
Alexandre Duplessis
Serge Belongie
FAtt
42
0
0
10 Oct 2024
Regulating Model Reliance on Non-Robust Features by Smoothing Input Marginal Density
Peiyu Yang
Naveed Akhtar
Mubarak Shah
Ajmal Saeed Mian
AAML
23
1
0
05 Jul 2024
Backdoor-based Explainable AI Benchmark for High Fidelity Evaluation of Attribution Methods
Peiyu Yang
Naveed Akhtar
Jiantong Jiang
Ajmal Saeed Mian
XAI
30
2
0
02 May 2024
Persistent Classification: A New Approach to Stability of Data and Adversarial Examples
Brian Bell
Michael Geyer
David Glickenstein
Keaton Hamm
C. Scheidegger
Amanda S. Fernandez
Juston Moore
AAML
36
0
0
11 Apr 2024
Structured Gradient-based Interpretations via Norm-Regularized Adversarial Training
Shizhan Gong
Qi Dou
Farzan Farnia
FAtt
37
2
0
06 Apr 2024
On the Faithfulness of Vision Transformer Explanations
Junyi Wu
Weitai Kang
Hao Tang
Yuan Hong
Yan Yan
19
6
0
01 Apr 2024
What Sketch Explainability Really Means for Downstream Tasks
Hmrishav Bandyopadhyay
Pinaki Nath Chowdhury
A. Bhunia
Aneeshan Sain
Tao Xiang
Yi-Zhe Song
30
4
0
14 Mar 2024
3VL: Using Trees to Improve Vision-Language Models' Interpretability
Nir Yellinek
Leonid Karlinsky
Raja Giryes
CoGe
VLM
49
4
0
28 Dec 2023
Keep the Faith: Faithful Explanations in Convolutional Neural Networks for Case-Based Reasoning
Tom Nuno Wolf
Fabian Bongratz
Anne-Marie Rickmann
Sebastian Polsterl
Christian Wachinger
AAML
FAtt
40
6
0
15 Dec 2023
Explanation-based Training with Differentiable Insertion/Deletion Metric-aware Regularizers
Yuya Yoshikawa
Tomoharu Iwata
14
0
0
19 Oct 2023
On the Trade-offs between Adversarial Robustness and Actionable Explanations
Satyapriya Krishna
Chirag Agarwal
Himabindu Lakkaraju
AAML
36
0
0
28 Sep 2023
Interpretability-Aware Vision Transformer
Yao Qiang
Chengyin Li
Prashant Khanduri
D. Zhu
ViT
80
7
0
14 Sep 2023
An Exact Kernel Equivalence for Finite Classification Models
Brian Bell
Michaela Geyer
David Glickenstein
Amanda Fernandez
Juston Moore
19
2
0
01 Aug 2023
Discriminative Feature Attributions: Bridging Post Hoc Explainability and Inherent Interpretability
Usha Bhalla
Suraj Srinivas
Himabindu Lakkaraju
FAtt
CML
26
6
0
27 Jul 2023
Contextual Reliability: When Different Features Matter in Different Contexts
Gaurav R. Ghosal
Amrith Rajagopal Setlur
Daniel S. Brown
Anca Dragan
Aditi Raghunathan
25
1
0
19 Jul 2023
B-cos Alignment for Inherently Interpretable CNNs and Vision Transformers
Moritz D Boehle
Navdeeppal Singh
Mario Fritz
Bernt Schiele
54
27
0
19 Jun 2023
Protein Design with Guided Discrete Diffusion
Nate Gruver
Samuel Stanton
Nathan C. Frey
Tim G. J. Rudner
I. Hotzel
J. Lafrance-Vanasse
A. Rajpal
Kyunghyun Cho
A. Wilson
DiffM
31
101
0
31 May 2023
Which Models have Perceptually-Aligned Gradients? An Explanation via Off-Manifold Robustness
Suraj Srinivas
Sebastian Bordt
Hima Lakkaraju
AAML
25
11
0
30 May 2023
On the Impact of Knowledge Distillation for Model Interpretability
Hyeongrok Han
Siwon Kim
Hyun-Soo Choi
Sungroh Yoon
10
4
0
25 May 2023
Better Understanding Differences in Attribution Methods via Systematic Evaluations
Sukrut Rao
Moritz D Boehle
Bernt Schiele
XAI
27
2
0
21 Mar 2023
Variational Information Pursuit for Interpretable Predictions
Aditya Chattopadhyay
Kwan Ho Ryan Chan
B. Haeffele
D. Geman
René Vidal
DRL
15
10
0
06 Feb 2023
Negative Flux Aggregation to Estimate Feature Attributions
X. Li
Deng Pan
Chengyin Li
Yao Qiang
D. Zhu
FAtt
6
6
0
17 Jan 2023
Post hoc Explanations may be Ineffective for Detecting Unknown Spurious Correlation
Julius Adebayo
M. Muelly
H. Abelson
Been Kim
16
86
0
09 Dec 2022
ModelDiff: A Framework for Comparing Learning Algorithms
Harshay Shah
Sung Min Park
Andrew Ilyas
A. Madry
SyDa
46
26
0
22 Nov 2022
A Rigorous Study Of The Deep Taylor Decomposition
Leon Sixt
Tim Landgraf
FAtt
AAML
17
4
0
14 Nov 2022
Interpretations Steered Network Pruning via Amortized Inferred Saliency Maps
Alireza Ganjdanesh
Shangqian Gao
Heng-Chiao Huang
FAtt
AAML
17
19
0
07 Sep 2022
SHAP-XRT: The Shapley Value Meets Conditional Independence Testing
Jacopo Teneggi
Beepul Bharti
Yaniv Romano
Jeremias Sulam
FAtt
20
3
0
14 Jul 2022
Interpretable by Design: Learning Predictors by Composing Interpretable Queries
Aditya Chattopadhyay
Stewart Slocum
B. Haeffele
René Vidal
D. Geman
24
21
0
03 Jul 2022
When are Post-hoc Conceptual Explanations Identifiable?
Tobias Leemann
Michael Kirchhof
Yao Rong
Enkelejda Kasneci
Gjergji Kasneci
50
10
0
28 Jun 2022
The Manifold Hypothesis for Gradient-Based Explanations
Sebastian Bordt
Uddeshya Upadhyay
Zeynep Akata
U. V. Luxburg
FAtt
AAML
16
12
0
15 Jun 2022
On the explainable properties of 1-Lipschitz Neural Networks: An Optimal Transport Perspective
M. Serrurier
Franck Mamalet
Thomas Fel
Louis Bethune
Thibaut Boissin
AAML
FAtt
22
4
0
14 Jun 2022
Towards Better Understanding Attribution Methods
Sukrut Rao
Moritz Bohle
Bernt Schiele
XAI
18
32
0
20 May 2022
B-cos Networks: Alignment is All We Need for Interpretability
Moritz D Boehle
Mario Fritz
Bernt Schiele
31
84
0
20 May 2022
A Unified Study of Machine Learning Explanation Evaluation Metrics
Yipei Wang
Xiaoqian Wang
XAI
14
7
0
27 Mar 2022
A Consistent and Efficient Evaluation Strategy for Attribution Methods
Yao Rong
Tobias Leemann
V. Borisov
Gjergji Kasneci
Enkelejda Kasneci
FAtt
23
92
0
01 Feb 2022
Sanity Simulations for Saliency Methods
Joon Sik Kim
Gregory Plumb
Ameet Talwalkar
FAtt
30
17
0
13 May 2021
Fast Hierarchical Games for Image Explanations
Jacopo Teneggi
Alexandre Luster
Jeremias Sulam
FAtt
26
17
0
13 Apr 2021
On Translation Invariance in CNNs: Convolutional Layers can Exploit Absolute Spatial Location
O. Kayhan
J. C. V. Gemert
209
232
0
16 Mar 2020
Adversarial examples from computational constraints
Sébastien Bubeck
Eric Price
Ilya P. Razenshteyn
AAML
62
230
0
25 May 2018
Trainability and Accuracy of Neural Networks: An Interacting Particle System Approach
Grant M. Rotskoff
Eric Vanden-Eijnden
59
118
0
02 May 2018
1