KR2025Proceedings of the 22nd International Conference on Principles of Knowledge Representation and ReasoningProceedings of the 22nd International Conference on Principles of Knowledge Representation and Reasoning

Melbourne, Australia. November 11-17, 2025.

Edited by

ISSN: 2334-1033
ISBN: 978-1-956792-08-9

Sponsored by
Published by

Copyright © 2025 International Joint Conferences on Artificial Intelligence Organization

Categorical Explaining Functors: Ensuring Coherence in Logical Explanations

  1. Stefano Fioravanti(University of Padua, University of Siena, Charles University)
  2. Francesco Giannini(Scuola Normale Superiore)
  3. Pietro Barbiero(IBM Research Zurich)
  4. Paolo Frazzetto(University of Padua)
  5. Roberto Confalonieri(University of Padua)
  6. Fabio Zanasi(University College London)
  7. Nicolò Navarin(University of Padua)

Keywords

  1. Explainable AI
  2. XAI
  3. Category Theory
  4. Fuzzy Explainers
  5. Logic Based Explanations
  6. Concept Bottleneck Models

Abstract

Post-hoc methods in Explainable AI (XAI) elucidate black-box models by identifying input features critical to the model's decision-making. Recent advancements in these methods have facilitated the generation of logic-based explanations that capture interactions among input features. However, these techniques often encounter critical limitations, notably the inability to ensure logical consistency and fidelity between generated explanations and the model's actual decision-making processes. Such inconsistencies jeopardize the reliability of explanations particularly in high-risk domains.

To address this gap, we introduce a novel, theoretically rigorous approach rooted in category theory. Specifically, we propose the concept of an explaining functor, which preserves logical entailment structurally between the explanations and the decisions of black-box models. By establishing a categorical framework, our method guarantees the coherence and accuracy of extracted explanations, thus overcoming the common pitfalls associated with heuristic-based explanation methods. We demonstrate the practical efficacy of our theoretical contributions through two synthetic benchmarks that highlight significant reductions in contradictory and unfaithful explanations. Our experiments show how our framework can provide mathematically grounded, compositional, and coherent explanations.