JOURNAL ARTICLE

Beyond Post-Hoc: Generating Inherently Verifiable Natural Language Explanations from AI Models

Revista, ZenIA, 10

Year: 2025 Journal:   Zenodo (CERN European Organization for Nuclear Research)   Publisher: European Organization for Nuclear Research

Abstract

The increasing complexity and widespread deployment of Artificial Intelligence (AI) models, particularly deep learning systems, have amplified the demand for explainability. Traditional Explainable AI (XAI) methods often rely on post-hoc approaches, generating explanations after a model has made a prediction. While valuable, these post-hoc explanations can suffer from issues of fidelity, consistency, and a lack of direct verifiability against the model's true decision-making process. This paper proposes a paradigm shift towards generating inherently verifiable natural language explanations. We argue for the integration of symbolic reasoning and formal verification techniques directly into AI model architectures, enabling systems to produce explanations that are not merely plausible but are demonstrably grounded in the model's internal logic. Such an approach aims to foster greater trust, accountability, and reliability in AI systems, especially in high-stakes domains where erroneous or unexplainable decisions can have severe consequences. We discuss a conceptual framework for constructing such models, outline the methodological challenges, and highlight the potential for hybrid neuro-symbolic AI to bridge the gap between high performance and verifiable transparency.

Keywords:
Verifiable secret sharing Natural language Bridge (graph theory) Natural (archaeology) Software deployment Key (lock) Reliability (semiconductor) Model checking

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
0
Refs
0.85
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Explainable Artificial Intelligence (XAI)
Physical Sciences →  Computer Science →  Artificial Intelligence
Adversarial Robustness in Machine Learning
Physical Sciences →  Computer Science →  Artificial Intelligence
Multimodal Machine Learning Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition

Related Documents

JOURNAL ARTICLE

Beyond Post-Hoc: Generating Inherently Verifiable Natural Language Explanations from AI Models

Revista, ZenIA, 10

Journal:   Zenodo (CERN European Organization for Nuclear Research) Year: 2025
JOURNAL ARTICLE

Interpretable AI: Beyond Post-Hoc Explanations Towards Inherently Understandable Models

Revista, ZenIA, 10

Journal:   Zenodo (CERN European Organization for Nuclear Research) Year: 2025
JOURNAL ARTICLE

Interpretable AI: Beyond Post-Hoc Explanations Towards Inherently Understandable Models

Revista, ZenIA, 10

Journal:   Zenodo (CERN European Organization for Nuclear Research) Year: 2025
BOOK-CHAPTER

Generating Natural Language Explanations from Plans

Chris Mellish

The MIT Press eBooks Year: 1990 Pages: 181-224
© 2026 ScienceGate Book Chapters — All rights reserved.