ABE: A Unified Framework for Robust and Faithful Attribution-Based Explainability

Attribution algorithms are essential for enhancing the interpretability and trustworthiness of deep learning models by identifying key features driving model decisions. Existing frameworks, such as InterpretDL and OmniXAI, integrate multiple attribution methods but suffer from scalability limitations, high coupling, theoretical constraints, and lack of user-friendly implementations, hindering neural network transparency and interoperability. To address these challenges, we propose Attribution-Based Explainability (ABE), a unified framework that formalizes Fundamental Attribution Methods and integrates state-of-the-art attribution algorithms while ensuring compliance with attribution axioms. ABE enables researchers to develop novel attribution techniques and enhances interpretability through four customizable modules: Robustness, Interpretability, Validation, and Data & Model. This framework provides a scalable, extensible foundation for advancing attribution-based explainability and fostering transparent AI systems. Our code is available at:this https URL.
View on arXiv@article{zhu2025_2505.06258, title={ ABE: A Unified Framework for Robust and Faithful Attribution-Based Explainability }, author={ Zhiyu Zhu and Jiayu Zhang and Zhibo Jin and Fang Chen and Jianlong Zhou }, journal={arXiv preprint arXiv:2505.06258}, year={ 2025 } }