Symbolic Methods Promoted for Rigorous Feature Attribution in XAI
Paper critiques non-rigorous non-symbolic XAI methods like SHAP and synthesizes symbolic alternatives for provable feature attribution to meet regulatory and enterprise accountability needs.
Non-symbolic explanation methods have dominated XAI for a decade but lack rigor required for high-stakes ML decisions (Huang, arXiv:2604.15898). The paper identifies provable shortcomings in Shapley value applications such as SHAP that can mislead human decision makers and overviews symbolic alternatives for provably correct relative feature importance (Huang, arXiv:2604.15898; Lundberg and Lee, arXiv:1705.07874).
Original coverage of SHAP-based tools has missed their axiomatic inconsistencies when applied to certain model classes and has under-emphasized alignment with regulatory demands such as the EU AI Act's transparency requirements for high-risk systems. Related formal XAI literature demonstrates that symbolic techniques using satisfiability solvers deliver exact explanations with computational trade-offs, connections the primary source only partially develops (Ignatiev et al., arXiv:1811.10652).
Patterns from prior XAI shifts show that usability-focused methods repeatedly encounter reproducibility failures under distribution shift; synthesis of the cited sources indicates symbolic rigor directly addresses the accountability gap for enterprises and regulators as opaque models scale, supplying provable guarantees absent in prevailing non-symbolic practice.
AXIOM: Popular tools like SHAP lack mathematical rigor and can mislead in high-stakes settings. Symbolic feature attribution methods supply the formal guarantees regulators and enterprises require as models grow more opaque.
Sources (3)
- [1]Towards Rigorous Explainability by Feature Attribution(https://arxiv.org/abs/2604.15898)
- [2]A Unified Approach to Interpreting Model Predictions(https://arxiv.org/abs/1705.07874)
- [3]Abduction-Based Explanations for Machine Learning Models(https://arxiv.org/abs/1811.10652)