A Validation Methodology for XAI Decision Support Systems Against Relational Domain Properties

ABSTRACT The global adoption of artificial intelligence (AI) has increased dramatically in recent years, becoming commonplace in many fields. Such a pervasiveness has led to changes in how AI is perceived, strengthening discussions on its societal consequences. Thus, a new class of requirements for...

Full description

Saved in:
Bibliographic Details
Published in:Journal of software : evolution and process Vol. 37; no. 10
Main Authors: De Angelis, Emanuele, De Angelis, Guglielmo, Mongelli, Maurizio, Proietti, Maurizio
Format: Journal Article
Language:English
Published: Chichester Wiley Subscription Services, Inc 01.10.2025
Subjects:
ISSN:2047-7473, 2047-7481
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:ABSTRACT The global adoption of artificial intelligence (AI) has increased dramatically in recent years, becoming commonplace in many fields. Such a pervasiveness has led to changes in how AI is perceived, strengthening discussions on its societal consequences. Thus, a new class of requirements for AI‐based solutions emerged. Broadly speaking, those on “explainability” aim to provide a transparent representation of the (often opaque) reasoning method that an AI‐based solution uses when prompted. This work presents a methodology for validating a class of explainable AI (XAI) models, called deterministic rule‐based models, which are used for expressing an explainable approximation of classifiers based on machine learning. The validation methodology combines logical deduction with constraint‐based reasoning in numerical domains, and it either succeeds or returns quantitative estimations of the invalid deviations found. This information allows us to assess the correctness of an XAI model, or in the case of deviations, to evaluate if it still can be deemed acceptable. The validation methodology has been applied to a simulation‐based study where the decision‐making process copes with the spread of SARS‐COV‐2 inside a railway station. The considered case study is a controlled but nontrivial example that shows the overall applicability of the methodology.
Bibliography:This work was supported by the project OPENNESS (N. A0375‐2020‐36616) funded by POR FESR LAZIO 2014‐2020 – GRUPPI DI RICERCA 2020, Future Artificial Intelligence Research (FAIR) (N. PE0000013) funded by Italian Recovery and Resilience Plan, and Gruppo Nazionale per il Calcolo Scientifico INdAM.
Funding
ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:2047-7473
2047-7481
DOI:10.1002/smr.70054