ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1806.07538
  4. Cited By
Towards Robust Interpretability with Self-Explaining Neural Networks

Towards Robust Interpretability with Self-Explaining Neural Networks

20 June 2018
David Alvarez-Melis
Tommi Jaakkola
    MILM
    XAI
ArXivPDFHTML

Papers citing "Towards Robust Interpretability with Self-Explaining Neural Networks"

50 / 507 papers shown
Title
What's meant by explainable model: A Scoping Review
What's meant by explainable model: A Scoping Review
Mallika Mainali
Rosina O. Weber
XAI
39
0
0
18 Jul 2023
SHAMSUL: Systematic Holistic Analysis to investigate Medical
  Significance Utilizing Local interpretability methods in deep learning for
  chest radiography pathology prediction
SHAMSUL: Systematic Holistic Analysis to investigate Medical Significance Utilizing Local interpretability methods in deep learning for chest radiography pathology prediction
Mahbub Ul Alam
Jaakko Hollmén
Jón R. Baldvinsson
R. Rahmani
FAtt
31
1
0
16 Jul 2023
Robust Ranking Explanations
Robust Ranking Explanations
Chao Chen
Chenghua Guo
Guixiang Ma
Ming Zeng
Xi Zhang
Sihong Xie
FAtt
AAML
35
0
0
08 Jul 2023
SHARCS: Shared Concept Space for Explainable Multimodal Learning
SHARCS: Shared Concept Space for Explainable Multimodal Learning
Gabriele Dominici
Pietro Barbiero
Lucie Charlotte Magister
Pietro Lio'
Nikola Simidjievski
36
5
0
01 Jul 2023
Towards Trustworthy Explanation: On Causal Rationalization
Towards Trustworthy Explanation: On Causal Rationalization
Wenbo Zhang
Tong Wu
Yunlong Wang
Yong Cai
Hengrui Cai
CML
21
18
0
25 Jun 2023
ProtoGate: Prototype-based Neural Networks with Global-to-local Feature
  Selection for Tabular Biomedical Data
ProtoGate: Prototype-based Neural Networks with Global-to-local Feature Selection for Tabular Biomedical Data
Xiangjian Jiang
Andrei Margeloiu
Nikola Simidjievski
M. Jamnik
OOD
34
10
0
21 Jun 2023
Selective Concept Models: Permitting Stakeholder Customisation at
  Test-Time
Selective Concept Models: Permitting Stakeholder Customisation at Test-Time
Matthew Barker
Katherine M. Collins
Krishnamurthy Dvijotham
Adrian Weller
Umang Bhatt
22
2
0
14 Jun 2023
Adversarial Attacks on the Interpretation of Neuron Activation
  Maximization
Adversarial Attacks on the Interpretation of Neuron Activation Maximization
Géraldin Nanfack
A. Fulleringer
Jonathan Marty
Michael Eickenberg
Eugene Belilovsky
AAML
FAtt
30
10
0
12 Jun 2023
Efficient GNN Explanation via Learning Removal-based Attribution
Efficient GNN Explanation via Learning Removal-based Attribution
Yao Rong
Guanchu Wang
Qizhang Feng
Ninghao Liu
Zirui Liu
Enkelejda Kasneci
Xia Hu
25
9
0
09 Jun 2023
Interpretable Deep Clustering for Tabular Data
Interpretable Deep Clustering for Tabular Data
Jonathan Svirsky
Ofir Lindenbaum
36
6
0
07 Jun 2023
Adversarial attacks and defenses in explainable artificial intelligence:
  A survey
Adversarial attacks and defenses in explainable artificial intelligence: A survey
Hubert Baniecki
P. Biecek
AAML
42
63
0
06 Jun 2023
Few Shot Rationale Generation using Self-Training with Dual Teachers
Few Shot Rationale Generation using Self-Training with Dual Teachers
Aditya Srikanth Veerubhotla
Lahari Poddar
J. Yin
Gyuri Szarvas
S. Eswaran
LRM
18
2
0
05 Jun 2023
Probabilistic Concept Bottleneck Models
Probabilistic Concept Bottleneck Models
Eunji Kim
Dahuin Jung
Sangha Park
Siwon Kim
Sung-Hoon Yoon
6
64
0
02 Jun 2023
An Experimental Investigation into the Evaluation of Explainability
  Methods
An Experimental Investigation into the Evaluation of Explainability Methods
Sédrick Stassin
A. Englebert
Géraldin Nanfack
Julien Albert
Nassim Versbraegen
Gilles Peiffer
Miriam Doh
Nicolas Riche
Benoit Frénay
Christophe De Vleeschouwer
XAI
ELM
16
5
0
25 May 2023
Concept-Centric Transformers: Enhancing Model Interpretability through
  Object-Centric Concept Learning within a Shared Global Workspace
Concept-Centric Transformers: Enhancing Model Interpretability through Object-Centric Concept Learning within a Shared Global Workspace
Jinyung Hong
Keun Hee Park
Theodore P. Pavlic
29
5
0
25 May 2023
On the Impact of Knowledge Distillation for Model Interpretability
On the Impact of Knowledge Distillation for Model Interpretability
Hyeongrok Han
Siwon Kim
Hyun-Soo Choi
Sungroh Yoon
24
4
0
25 May 2023
Causal Analysis for Robust Interpretability of Neural Networks
Causal Analysis for Robust Interpretability of Neural Networks
Ola Ahmad
Nicolas Béreux
Loïc Baret
V. Hashemi
Freddy Lecue
CML
23
3
0
15 May 2023
Tackling Interpretability in Audio Classification Networks with
  Non-negative Matrix Factorization
Tackling Interpretability in Audio Classification Networks with Non-negative Matrix Factorization
Jayneel Parekh
Sanjeel Parekh
Pavlo Mozharovskyi
Gaël Richard
Florence dÁlché-Buc
33
6
0
11 May 2023
Understanding cirrus clouds using explainable machine learning
Understanding cirrus clouds using explainable machine learning
K. Jeggle
D. Neubauer
Gustau Camps-Valls
Ulrike Lohmann
21
3
0
03 May 2023
Discover and Cure: Concept-aware Mitigation of Spurious Correlation
Discover and Cure: Concept-aware Mitigation of Spurious Correlation
Shirley Wu
Mert Yuksekgonul
Linjun Zhang
James Zou
80
56
0
01 May 2023
Learning Bottleneck Concepts in Image Classification
Learning Bottleneck Concepts in Image Classification
Bowen Wang
Liangzhi Li
Yuta Nakashima
Hajime Nagahara
SSL
25
42
0
20 Apr 2023
A Brief Review of Explainable Artificial Intelligence in Healthcare
A Brief Review of Explainable Artificial Intelligence in Healthcare
Zahra Sadeghi
R. Alizadehsani
M. Cifci
Samina Kausar
Rizwan Rehman
...
A. Shoeibi
H. Moosaei
Milan Hladík
Saeid Nahavandi
P. Pardalos
19
13
0
04 Apr 2023
Fine-tuning of explainable CNNs for skin lesion classification based on
  dermatologists' feedback towards increasing trust
Fine-tuning of explainable CNNs for skin lesion classification based on dermatologists' feedback towards increasing trust
Md Abdul Kadir
Fabrizio Nunnari
Daniel Sonntag
FAtt
11
1
0
03 Apr 2023
Posthoc Interpretation via Quantization
Posthoc Interpretation via Quantization
Francesco Paissan
Cem Subakan
Mirco Ravanelli
MQ
21
6
0
22 Mar 2023
Causal Discovery from Temporal Data: An Overview and New Perspectives
Causal Discovery from Temporal Data: An Overview and New Perspectives
Chang Gong
Di Yao
Chuzhe Zhang
Wenbin Li
Jingping Bi
AI4TS
CML
16
17
0
17 Mar 2023
Beyond Demographic Parity: Redefining Equal Treatment
Beyond Demographic Parity: Redefining Equal Treatment
Carlos Mougan
Laura State
Antonio Ferrara
Salvatore Ruggieri
Steffen Staab
FaML
33
1
0
14 Mar 2023
ICICLE: Interpretable Class Incremental Continual Learning
ICICLE: Interpretable Class Incremental Continual Learning
Dawid Rymarczyk
Joost van de Weijer
Bartosz Zieliñski
Bartlomiej Twardowski
CLL
32
28
0
14 Mar 2023
A Test Statistic Estimation-based Approach for Establishing
  Self-interpretable CNN-based Binary Classifiers
A Test Statistic Estimation-based Approach for Establishing Self-interpretable CNN-based Binary Classifiers
S. Sengupta
M. Anastasio
MedIm
33
6
0
13 Mar 2023
On the Benefits of Biophysical Synapses
On the Benefits of Biophysical Synapses
Julian Lemmel
Radu Grosu
9
0
0
08 Mar 2023
A Multi-Grained Self-Interpretable Symbolic-Neural Model For
  Single/Multi-Labeled Text Classification
A Multi-Grained Self-Interpretable Symbolic-Neural Model For Single/Multi-Labeled Text Classification
Xiang Hu
Xinyu Kong
Kewei Tu
MILM
BDL
23
5
0
06 Mar 2023
Finding the right XAI method -- A Guide for the Evaluation and Ranking
  of Explainable AI Methods in Climate Science
Finding the right XAI method -- A Guide for the Evaluation and Ranking of Explainable AI Methods in Climate Science
P. Bommer
M. Kretschmer
Anna Hedström
Dilyara Bareeva
Marina M.-C. Höhne
46
38
0
01 Mar 2023
Inherently Interpretable Multi-Label Classification Using Class-Specific
  Counterfactuals
Inherently Interpretable Multi-Label Classification Using Class-Specific Counterfactuals
Susu Sun
S. Woerner
Andreas Maier
Lisa M. Koch
Christian F. Baumgartner
FAtt
35
16
0
01 Mar 2023
A Closer Look at the Intervention Procedure of Concept Bottleneck Models
A Closer Look at the Intervention Procedure of Concept Bottleneck Models
Sungbin Shin
Yohan Jo
Sungsoo Ahn
Namhoon Lee
23
30
0
28 Feb 2023
Function Composition in Trustworthy Machine Learning: Implementation
  Choices, Insights, and Questions
Function Composition in Trustworthy Machine Learning: Implementation Choices, Insights, and Questions
Manish Nagireddy
Moninder Singh
Samuel C. Hoffman
Evaline Ju
K. Ramamurthy
Kush R. Varshney
30
1
0
17 Feb 2023
On marginal feature attributions of tree-based models
On marginal feature attributions of tree-based models
Khashayar Filom
A. Miroshnikov
Konstandinos Kotsiopoulos
Arjun Ravi Kannan
FAtt
22
3
0
16 Feb 2023
The Meta-Evaluation Problem in Explainable AI: Identifying Reliable
  Estimators with MetaQuantus
The Meta-Evaluation Problem in Explainable AI: Identifying Reliable Estimators with MetaQuantus
Anna Hedström
P. Bommer
Kristoffer K. Wickstrom
Wojciech Samek
Sebastian Lapuschkin
Marina M.-C. Höhne
37
21
0
14 Feb 2023
A novel approach to generate datasets with XAI ground truth to evaluate
  image models
A novel approach to generate datasets with XAI ground truth to evaluate image models
Miquel Miró-Nicolau
Antoni Jaume-i-Capó
Gabriel Moyà Alcover
22
4
0
11 Feb 2023
Symbolic Metamodels for Interpreting Black-boxes Using Primitive
  Functions
Symbolic Metamodels for Interpreting Black-boxes Using Primitive Functions
Mahed Abroshan
Saumitra Mishra
Mohammad Mahdi Khalili
33
4
0
09 Feb 2023
Variational Information Pursuit for Interpretable Predictions
Variational Information Pursuit for Interpretable Predictions
Aditya Chattopadhyay
Kwan Ho Ryan Chan
B. Haeffele
D. Geman
René Vidal
DRL
21
10
0
06 Feb 2023
Improving Interpretability via Explicit Word Interaction Graph Layer
Improving Interpretability via Explicit Word Interaction Graph Layer
Arshdeep Sekhon
Hanjie Chen
A. Shrivastava
Zhe Wang
Yangfeng Ji
Yanjun Qi
AI4CE
MILM
25
6
0
03 Feb 2023
Robust online active learning
Robust online active learning
Davide Cacciarelli
M. Kulahci
J. Tyssedal
16
12
0
01 Feb 2023
A Survey of Explainable AI in Deep Visual Modeling: Methods and Metrics
A Survey of Explainable AI in Deep Visual Modeling: Methods and Metrics
Naveed Akhtar
XAI
VLM
32
7
0
31 Jan 2023
ProtoSeg: Interpretable Semantic Segmentation with Prototypical Parts
ProtoSeg: Interpretable Semantic Segmentation with Prototypical Parts
Mikolaj Sacha
Dawid Rymarczyk
Lukasz Struski
Jacek Tabor
Bartosz Zieliñski
VLM
35
29
0
28 Jan 2023
Towards Robust Metrics for Concept Representation Evaluation
Towards Robust Metrics for Concept Representation Evaluation
M. Zarlenga
Pietro Barbiero
Z. Shams
Dmitry Kazhdan
Umang Bhatt
Adrian Weller
M. Jamnik
21
24
0
25 Jan 2023
Interpretability in Activation Space Analysis of Transformers: A Focused
  Survey
Interpretability in Activation Space Analysis of Transformers: A Focused Survey
Soniya Vijayakumar
AI4CE
35
3
0
22 Jan 2023
Holistically Explainable Vision Transformers
Holistically Explainable Vision Transformers
Moritz D Boehle
Mario Fritz
Bernt Schiele
ViT
38
9
0
20 Jan 2023
Sanity checks and improvements for patch visualisation in
  prototype-based image classification
Sanity checks and improvements for patch visualisation in prototype-based image classification
Romain Xu-Darme
Georges Quénot
Zakaria Chihani
M. Rousset
10
3
0
20 Jan 2023
Mortality Prediction with Adaptive Feature Importance Recalibration for
  Peritoneal Dialysis Patients: a deep-learning-based study on a real-world
  longitudinal follow-up dataset
Mortality Prediction with Adaptive Feature Importance Recalibration for Peritoneal Dialysis Patients: a deep-learning-based study on a real-world longitudinal follow-up dataset
Liantao Ma
Chaohe Zhang
Junyi Gao
Xianfeng Jiao
Zhihao Yu
...
Yasha Wang
Wen Tang
Xinju Zhao
Wenjie Ruan
Tao Wang
OOD
11
2
0
17 Jan 2023
Img2Tab: Automatic Class Relevant Concept Discovery from StyleGAN
  Features for Explainable Image Classification
Img2Tab: Automatic Class Relevant Concept Discovery from StyleGAN Features for Explainable Image Classification
Y. Song
S. K. Shyn
Kwang-su Kim
VLM
21
5
0
16 Jan 2023
Learning Support and Trivial Prototypes for Interpretable Image
  Classification
Learning Support and Trivial Prototypes for Interpretable Image Classification
Chong Wang
Yuyuan Liu
Yuanhong Chen
Fengbei Liu
Yu Tian
Davis J. McCarthy
Helen Frazer
G. Carneiro
34
24
0
08 Jan 2023
Previous
12345...91011
Next