pith. machine review for the scientific record. sign in

arxiv: 2602.04120 · v2 · submitted 2026-02-04 · 💻 cs.LG · cs.AI· cs.DC· cs.SE

Recognition: unknown

Scalable Explainability-as-a-Service (XaaS) for Edge AI Systems

Authors on Pith no claims yet
classification 💻 cs.LG cs.AIcs.DCcs.SE
keywords explanationxaasedgeacrossexplanationslatencysystemsthree
0
0 comments X
read the original abstract

Though Explainable AI (XAI) has made significant advancements, its inclusion in edge and IoT systems is typically ad-hoc and inefficient. Most current methods are "coupled" in such a way that they generate explanations simultaneously with model inferences. As a result, these approaches incur redundant computation, high latency and poor scalability when deployed across heterogeneous sets of edge devices. In this work we propose Explainability-as-a-Service (XaaS), a distributed architecture for treating explainability as a first-class system service (as opposed to a model-specific feature). The key innovation in our proposed XaaS architecture is that it decouples inference from explanation generation allowing edge devices to request, cache and verify explanations subject to resource and latency constraints. To achieve this, we introduce three main innovations: (1) A distributed explanation cache with a semantic similarity based explanation retrieval method which significantly reduces redundant computation; (2) A lightweight verification protocol that ensures the fidelity of both cached and newly generated explanations; and (3) An adaptive explanation engine that chooses explanation methods based upon device capability and user requirement. We evaluated the performance of XaaS on three real-world edge-AI use cases: (i) manufacturing quality control; (ii) autonomous vehicle perception; and (iii) healthcare diagnostics. Experimental results show that XaaS reduces latency by 38% while maintaining high explanation quality across three real-world deployments. Overall, this work enables the deployment of transparent and accountable AI across large scale, heterogeneous IoT systems, and bridges the gap between XAI research and edge-practicality.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 1 Pith paper

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Persistent and Conversational Multi-Method Explainability for Trustworthy Financial AI

    cs.AI 2026-05 unverdicted novelty 4.0

    An architecture stores XAI explanations persistently in searchable storage and uses RAG to synthesize multiple methods conversationally, cutting hallucination rates by 36% in a FinBERT financial sentiment demo.