ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1909.13584
  4. Cited By
Interpretations are useful: penalizing explanations to align neural
  networks with prior knowledge

Interpretations are useful: penalizing explanations to align neural networks with prior knowledge

30 September 2019
Laura Rieger
Chandan Singh
W. James Murdoch
Bin Yu
    FAtt
ArXivPDFHTML

Papers citing "Interpretations are useful: penalizing explanations to align neural networks with prior knowledge"

50 / 53 papers shown
Title
Explanation Regularisation through the Lens of Attributions
Explanation Regularisation through the Lens of Attributions
Pedro Ferreira
Wilker Aziz
Ivan Titov
46
1
0
23 Jul 2024
Exploring the Trade-off Between Model Performance and Explanation
  Plausibility of Text Classifiers Using Human Rationales
Exploring the Trade-off Between Model Performance and Explanation Plausibility of Text Classifiers Using Human Rationales
Lucas Resck
Marcos M. Raimundo
Jorge Poco
50
1
0
03 Apr 2024
Improving deep learning with prior knowledge and cognitive models: A
  survey on enhancing explainability, adversarial robustness and zero-shot
  learning
Improving deep learning with prior knowledge and cognitive models: A survey on enhancing explainability, adversarial robustness and zero-shot learning
F. Mumuni
A. Mumuni
AAML
37
5
0
11 Mar 2024
Identifying Spurious Correlations using Counterfactual Alignment
Identifying Spurious Correlations using Counterfactual Alignment
Joseph Paul Cohen
Louis Blankemeier
Akshay S. Chaudhari
CML
55
1
0
01 Dec 2023
Concept Distillation: Leveraging Human-Centered Explanations for Model
  Improvement
Concept Distillation: Leveraging Human-Centered Explanations for Model Improvement
Avani Gupta
Saurabh Saini
P. J. Narayanan
28
6
0
26 Nov 2023
Explaining black boxes with a SMILE: Statistical Model-agnostic
  Interpretability with Local Explanations
Explaining black boxes with a SMILE: Statistical Model-agnostic Interpretability with Local Explanations
Koorosh Aslansefat
Mojgan Hashemian
M. Walker
Mohammed Naveed Akram
Ioannis Sorokos
Y. Papadopoulos
FAtt
AAML
19
2
0
13 Nov 2023
Is Last Layer Re-Training Truly Sufficient for Robustness to Spurious
  Correlations?
Is Last Layer Re-Training Truly Sufficient for Robustness to Spurious Correlations?
Phuong Quynh Le
Jorg Schlotterer
Christin Seifert
OOD
16
6
0
01 Aug 2023
Mitigating Bias: Enhancing Image Classification by Improving Model
  Explanations
Mitigating Bias: Enhancing Image Classification by Improving Model Explanations
Raha Ahmadi
Mohammad Javad Rajabi
Mohammad Khalooiem
Mohammad Sabokrou
31
0
0
04 Jul 2023
One Explanation Does Not Fit XIL
One Explanation Does Not Fit XIL
Felix Friedrich
David Steinmann
Kristian Kersting
LRM
37
2
0
14 Apr 2023
Preemptively Pruning Clever-Hans Strategies in Deep Neural Networks
Preemptively Pruning Clever-Hans Strategies in Deep Neural Networks
Lorenz Linhardt
Klaus-Robert Muller
G. Montavon
AAML
26
7
0
12 Apr 2023
Are Data-driven Explanations Robust against Out-of-distribution Data?
Are Data-driven Explanations Robust against Out-of-distribution Data?
Tang Li
Fengchun Qiao
Mengmeng Ma
Xiangkai Peng
OODD
OOD
33
10
0
29 Mar 2023
Learning with Explanation Constraints
Learning with Explanation Constraints
Rattana Pukdee
Dylan Sam
J. Zico Kolter
Maria-Florina Balcan
Pradeep Ravikumar
FAtt
32
6
0
25 Mar 2023
Removing confounding information from fetal ultrasound images
Removing confounding information from fetal ultrasound images
K. Mikolaj
Manxi Lin
Zahra Bashir
M. B. S. Svendsen
M. Tolsgaard
Anders Christensen
Aasa Feragen
27
3
0
24 Mar 2023
Towards Learning and Explaining Indirect Causal Effects in Neural
  Networks
Towards Learning and Explaining Indirect Causal Effects in Neural Networks
Abbaavaram Gowtham Reddy
Saketh Bachu
Harsh Nilesh Pathak
Ben Godfrey
V. Balasubramanian
V. Varshaneya
Satya Narayanan Kar
CML
31
0
0
24 Mar 2023
Informing clinical assessment by contextualizing post-hoc explanations
  of risk prediction models in type-2 diabetes
Informing clinical assessment by contextualizing post-hoc explanations of risk prediction models in type-2 diabetes
Shruthi Chari
Prasanth Acharya
Daniel Gruen
Olivia R. Zhang
Elif Eyigoz
...
Oshani Seneviratne
Fernando Jose Suarez Saiz
Pablo Meyer
Prithwish Chakraborty
D. McGuinness
24
17
0
11 Feb 2023
On the Relationship Between Explanation and Prediction: A Causal View
On the Relationship Between Explanation and Prediction: A Causal View
Amir-Hossein Karimi
Krikamol Muandet
Simon Kornblith
Bernhard Schölkopf
Been Kim
FAtt
CML
34
14
0
13 Dec 2022
Going Beyond XAI: A Systematic Survey for Explanation-Guided Learning
Going Beyond XAI: A Systematic Survey for Explanation-Guided Learning
Yuyang Gao
Siyi Gu
Junji Jiang
S. Hong
Dazhou Yu
Liang Zhao
29
39
0
07 Dec 2022
Knowledge-augmented Deep Learning and Its Applications: A Survey
Knowledge-augmented Deep Learning and Its Applications: A Survey
Zijun Cui
Tian Gao
Kartik Talamadupula
Qiang Ji
27
18
0
30 Nov 2022
Calibration Meets Explanation: A Simple and Effective Approach for Model
  Confidence Estimates
Calibration Meets Explanation: A Simple and Effective Approach for Model Confidence Estimates
Dongfang Li
Baotian Hu
Qingcai Chen
16
8
0
06 Nov 2022
XMD: An End-to-End Framework for Interactive Explanation-Based Debugging
  of NLP Models
XMD: An End-to-End Framework for Interactive Explanation-Based Debugging of NLP Models
Dong-Ho Lee
Akshen Kadakia
Brihi Joshi
Aaron Chan
Ziyi Liu
...
Takashi Shibuya
Ryosuke Mitani
Toshiyuki Sekiya
Jay Pujara
Xiang Ren
LRM
40
9
0
30 Oct 2022
XC: Exploring Quantitative Use Cases for Explanations in 3D Object
  Detection
XC: Exploring Quantitative Use Cases for Explanations in 3D Object Detection
Sunsheng Gu
Vahdat Abdelzad
Krzysztof Czarnecki
19
1
0
20 Oct 2022
Using Language to Extend to Unseen Domains
Using Language to Extend to Unseen Domains
Lisa Dunlap
Clara Mohri
Devin Guillory
Han Zhang
Trevor Darrell
Joseph E. Gonzalez
Aditi Raghunanthan
Anja Rohrbach
VLM
20
35
0
18 Oct 2022
Fairness via Adversarial Attribute Neighbourhood Robust Learning
Fairness via Adversarial Attribute Neighbourhood Robust Learning
Q. Qi
Shervin Ardeshir
Yi Tian Xu
Tianbao Yang
40
0
0
12 Oct 2022
Goal Misgeneralization: Why Correct Specifications Aren't Enough For
  Correct Goals
Goal Misgeneralization: Why Correct Specifications Aren't Enough For Correct Goals
Rohin Shah
Vikrant Varma
Ramana Kumar
Mary Phuong
Victoria Krakovna
J. Uesato
Zachary Kenton
37
68
0
04 Oct 2022
Domain Classification-based Source-specific Term Penalization for Domain
  Adaptation in Hate-speech Detection
Domain Classification-based Source-specific Term Penalization for Domain Adaptation in Hate-speech Detection
Tulika Bose
Nikolaos Aletras
Irina Illina
Dominique Fohr
19
0
0
18 Sep 2022
Leveraging Explanations in Interactive Machine Learning: An Overview
Leveraging Explanations in Interactive Machine Learning: An Overview
Stefano Teso
Öznur Alkan
Wolfgang Stammer
Elizabeth M. Daly
XAI
FAtt
LRM
26
62
0
29 Jul 2022
Dynamically Refined Regularization for Improving Cross-corpora Hate
  Speech Detection
Dynamically Refined Regularization for Improving Cross-corpora Hate Speech Detection
Tulika Bose
Nikolaos Aletras
Irina Illina
Dominique Fohr
45
5
0
23 Mar 2022
FairPrune: Achieving Fairness Through Pruning for Dermatological Disease
  Diagnosis
FairPrune: Achieving Fairness Through Pruning for Dermatological Disease Diagnosis
Yawen Wu
Dewen Zeng
Xiaowei Xu
Yiyu Shi
Jingtong Hu
MedIm
29
51
0
04 Mar 2022
Learning Robust Convolutional Neural Networks with Relevant Feature
  Focusing via Explanations
Learning Robust Convolutional Neural Networks with Relevant Feature Focusing via Explanations
Kazuki Adachi
Shin'ya Yamaguchi
OOD
35
0
0
09 Feb 2022
Right for the Right Latent Factors: Debiasing Generative Models via
  Disentanglement
Right for the Right Latent Factors: Debiasing Generative Models via Disentanglement
Xiaoting Shao
Karl Stelzner
Kristian Kersting
CML
DRL
22
3
0
01 Feb 2022
Explainable Deep Learning in Healthcare: A Methodological Survey from an
  Attribution View
Explainable Deep Learning in Healthcare: A Methodological Survey from an Attribution View
Di Jin
Elena Sergeeva
W. Weng
Geeticka Chauhan
Peter Szolovits
OOD
33
55
0
05 Dec 2021
What to Learn, and How: Toward Effective Learning from Rationales
What to Learn, and How: Toward Effective Learning from Rationales
Samuel Carton
Surya Kanoria
Chenhao Tan
45
22
0
30 Nov 2021
Matching Learned Causal Effects of Neural Networks with Domain Priors
Matching Learned Causal Effects of Neural Networks with Domain Priors
Sai Srinivas Kancheti
Abbavaram Gowtham Reddy
V. Balasubramanian
Amit Sharma
CML
33
12
0
24 Nov 2021
Self-Interpretable Model with TransformationEquivariant Interpretation
Self-Interpretable Model with TransformationEquivariant Interpretation
Yipei Wang
Xiaoqian Wang
38
23
0
09 Nov 2021
Transparency of Deep Neural Networks for Medical Image Analysis: A
  Review of Interpretability Methods
Transparency of Deep Neural Networks for Medical Image Analysis: A Review of Interpretability Methods
Zohaib Salahuddin
Henry C. Woodruff
A. Chatterjee
Philippe Lambin
21
302
0
01 Nov 2021
SIM-ECG: A Signal Importance Mask-driven ECGClassification System
SIM-ECG: A Signal Importance Mask-driven ECGClassification System
K. Dharma
Chicheng Zhang
C. Gniady
P. Agarwal
Sushil Sharma
31
0
0
28 Oct 2021
Consistent Explanations by Contrastive Learning
Consistent Explanations by Contrastive Learning
Vipin Pillai
Soroush Abbasi Koohpayegani
Ashley Ouligian
Dennis Fong
Hamed Pirsiavash
FAtt
20
21
0
01 Oct 2021
Improving the trustworthiness of image classification models by
  utilizing bounding-box annotations
Improving the trustworthiness of image classification models by utilizing bounding-box annotations
K. Dharma
Chicheng Zhang
32
5
0
15 Aug 2021
Fairness via Representation Neutralization
Fairness via Representation Neutralization
Mengnan Du
Subhabrata Mukherjee
Guanchu Wang
Ruixiang Tang
Ahmed Hassan Awadallah
Xia Hu
25
78
0
23 Jun 2021
Towards Robust Classification Model by Counterfactual and Invariant Data
  Generation
Towards Robust Classification Model by Counterfactual and Invariant Data Generation
C. Chang
George Adam
Anna Goldenberg
OOD
CML
27
31
0
02 Jun 2021
Explainable Machine Learning with Prior Knowledge: An Overview
Explainable Machine Learning with Prior Knowledge: An Overview
Katharina Beckh
Sebastian Müller
Matthias Jakobs
Vanessa Toborek
Hanxiao Tan
Raphael Fischer
Pascal Welke
Sebastian Houben
Laura von Rueden
XAI
22
28
0
21 May 2021
Explanation-Based Human Debugging of NLP Models: A Survey
Explanation-Based Human Debugging of NLP Models: A Survey
Piyawat Lertvittayakumjorn
Francesca Toni
LRM
42
79
0
30 Apr 2021
Shapley Explanation Networks
Shapley Explanation Networks
Rui Wang
Xiaoqian Wang
David I. Inouye
TDI
FAtt
24
44
0
06 Apr 2021
No Subclass Left Behind: Fine-Grained Robustness in Coarse-Grained
  Classification Problems
No Subclass Left Behind: Fine-Grained Robustness in Coarse-Grained Classification Problems
N. Sohoni
Jared A. Dunnmon
Geoffrey Angus
Albert Gu
Christopher Ré
30
242
0
25 Nov 2020
Right for the Right Concept: Revising Neuro-Symbolic Concepts by
  Interacting with their Explanations
Right for the Right Concept: Revising Neuro-Symbolic Concepts by Interacting with their Explanations
Wolfgang Stammer
P. Schramowski
Kristian Kersting
FAtt
14
107
0
25 Nov 2020
Learning Variational Word Masks to Improve the Interpretability of
  Neural Text Classifiers
Learning Variational Word Masks to Improve the Interpretability of Neural Text Classifiers
Hanjie Chen
Yangfeng Ji
AAML
VLM
15
63
0
01 Oct 2020
Trustworthy Convolutional Neural Networks: A Gradient Penalized-based
  Approach
Trustworthy Convolutional Neural Networks: A Gradient Penalized-based Approach
Nicholas F Halliwell
Freddy Lecue
FAtt
25
9
0
29 Sep 2020
Widening the Pipeline in Human-Guided Reinforcement Learning with
  Explanation and Context-Aware Data Augmentation
Widening the Pipeline in Human-Guided Reinforcement Learning with Explanation and Context-Aware Data Augmentation
L. Guan
Mudit Verma
Sihang Guo
Ruohan Zhang
Subbarao Kambhampati
43
42
0
26 Jun 2020
Contextualizing Hate Speech Classifiers with Post-hoc Explanation
Contextualizing Hate Speech Classifiers with Post-hoc Explanation
Brendan Kennedy
Xisen Jin
Aida Mostafazadeh Davani
Morteza Dehghani
Xiang Ren
13
137
0
05 May 2020
Machine Learning in Python: Main developments and technology trends in
  data science, machine learning, and artificial intelligence
Machine Learning in Python: Main developments and technology trends in data science, machine learning, and artificial intelligence
S. Raschka
Joshua Patterson
Corey J. Nolet
AI4CE
24
483
0
12 Feb 2020
12
Next