Recognition: unknown
Decision Trace Schema for Governance Evidence in Real-Time Risk Systems
Pith reviewed 2026-05-10 16:41 UTC · model grok-4.3
The pith
The Decision Event Schema supplies a single JSON structure that records governance evidence across ML inference, rule evaluation, cross-system coupling, and metadata layers for each automated decision.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
The Decision Event Schema (DES) is a JSON Schema specification that bridges four infrastructure layers—ML inference, rule/policy evaluation, cross-system coupling, and governance metadata—within a single per-decision event structure. It employs degradation-aware field design where each of six top-level field groups maps to a governance evidence property and the degradation type it must resist, defines ten required root-level fields, and introduces a tiered evidence strategy (lightweight, sampled, full) that enables organizations to match evidence completeness to decision risk and throughput. A mechanism feasibility analysis demonstrates compatibility with the highest-throughput integrity at
What carries the argument
The Decision Event Schema (DES), a JSON Schema with ten required root-level fields grouped into six top-level categories each tied to a governance evidence property and its resistance to a specific degradation type, plus support for three evidence tiers.
If this is right
- Practitioners can adopt DES directly or extend it with namespace extensions for domain-specific needs.
- Regulators receive a clear mapping from legal requirements to minimum evidence tiers.
- The schema remains compatible with high-throughput integrity mechanisms used at production decision rates.
- No other format among the twenty-five-plus examined covers all four infrastructure layers simultaneously.
Where Pith is reading between the lines
- Widespread use could produce consistent audit trails across different automated decision platforms.
- Live-system tests would be needed to measure actual overhead and completeness under real risk loads.
- The tiered design could be adapted to match the specific evidence rules of individual regulations such as financial or data-protection standards.
- The same structure might later support emerging requirements for AI accountability that go beyond today's rules.
Load-bearing premise
That defining fields for the four layers plus degradation-aware design and tiered evidence is enough to solve the Fragmented Trace Problem without additional real-world validation or performance data.
What would settle it
Finding any existing format that already captures evidence from all four layers at once, or running a production deployment that shows DES cannot maintain required evidence integrity at scale.
read the original abstract
Automated decision systems produce operational data across multiple infrastructure layers, yet no single logging format captures the complete governance-relevant record of how a decision was reached. Regulatory frameworks prescribe what must be recorded without specifying a data model for how to record it -- a gap this paper terms the Fragmented Trace Problem. Following a design science methodology, the paper presents the Decision Event Schema (DES), a JSON Schema specification that bridges four infrastructure layers -- ML inference, rule/policy evaluation, cross-system coupling, and governance metadata -- within a single per-decision event structure. The schema employs degradation-aware field design: each of six top-level field groups maps to a governance evidence property and the degradation type it must resist. DES defines ten required root-level fields and introduces a tiered evidence strategy (lightweight, sampled, full) that enables organizations to match evidence completeness to decision risk and throughput. A mechanism feasibility analysis demonstrates compatibility with the highest-throughput integrity mechanisms at production-scale decision rates. Evaluation against 25+ existing formats confirms that DES is the only specification covering all four layers simultaneously. The schema offers practitioners a reference adoptable directly or adaptable through namespace extensions, and regulators a mapping from requirements to minimum evidence tiers.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper introduces the Decision Event Schema (DES), a JSON Schema specification to address the Fragmented Trace Problem in automated decision systems. Following design science methodology, DES unifies governance-relevant traces across four infrastructure layers (ML inference, rule/policy evaluation, cross-system coupling, and governance metadata) via degradation-aware field design, ten required root-level fields, and a tiered evidence strategy (lightweight, sampled, full). It includes a mechanism feasibility analysis for production-scale throughput and claims, based on comparison to 25+ existing formats, to be the only specification covering all four layers simultaneously, offering a reference for practitioners and regulators.
Significance. If the evaluation details and mappings are provided and the schema proves complete in practice, DES could fill a practical gap between regulatory requirements for decision traceability and implementable data models in high-stakes automated systems. The design science approach, degradation-aware design, and feasibility analysis for high-throughput integrity mechanisms are constructive contributions that could support adoption or adaptation via namespace extensions.
major comments (2)
- [Abstract and Evaluation] Abstract and Evaluation section: The central uniqueness claim ('DES is the only specification covering all four layers simultaneously') is load-bearing but unverifiable without an explicit definition of 'covering' (e.g., must a format implement specific fields for a layer or only conceptually address it?), the selection criteria for the 25+ formats, and a table or mapping showing per-format, per-layer coverage. This prevents assessment of completeness or potential omissions in the benchmark.
- [Feasibility analysis] Feasibility analysis and solution claim: The analysis addresses throughput compatibility but does not include any empirical completeness check, case study, or audit against real decision logs to confirm that the ten required fields plus six groups capture all governance-relevant traces. This leaves the assertion that DES plus tiered evidence and degradation-aware design solves the Fragmented Trace Problem without additional validation mechanisms untested.
minor comments (1)
- [Abstract] The abstract and summary would benefit from a brief parenthetical example of one degradation-aware field and its governance property to illustrate the design without requiring the full schema.
Simulated Author's Rebuttal
We thank the referee for the constructive and detailed feedback. The comments highlight areas where additional clarity and support will strengthen the manuscript. We address each major comment below and will incorporate revisions as noted.
read point-by-point responses
-
Referee: [Abstract and Evaluation] Abstract and Evaluation section: The central uniqueness claim ('DES is the only specification covering all four layers simultaneously') is load-bearing but unverifiable without an explicit definition of 'covering' (e.g., must a format implement specific fields for a layer or only conceptually address it?), the selection criteria for the 25+ formats, and a table or mapping showing per-format, per-layer coverage. This prevents assessment of completeness or potential omissions in the benchmark.
Authors: We agree that the uniqueness claim requires explicit supporting details to be verifiable. In the revised manuscript, we will add: (1) a precise definition of 'covering' (a format covers a layer when it includes dedicated fields or structures for recording evidence specific to that layer, rather than only conceptual mention); (2) the selection criteria used for the 25+ formats (production logging standards, open-source governance tools, ML observability formats, and regulatory-aligned schemas); and (3) a mapping table in the Evaluation section showing per-format coverage of the four layers with brief justifications. These additions will allow direct assessment of the claim and any potential gaps. revision: yes
-
Referee: [Feasibility analysis] Feasibility analysis and solution claim: The analysis addresses throughput compatibility but does not include any empirical completeness check, case study, or audit against real decision logs to confirm that the ten required fields plus six groups capture all governance-relevant traces. This leaves the assertion that DES plus tiered evidence and degradation-aware design solves the Fragmented Trace Problem without additional validation mechanisms untested.
Authors: The feasibility analysis focuses on analytical modeling of throughput and integrity mechanism compatibility at production scale. We acknowledge that the manuscript does not provide an empirical audit or case study against real decision logs, which would offer stronger confirmation of completeness in operational settings. As a design science paper, the core claim rests on the schema's construction to address the four layers by design. In revision, we will add an illustrative case study using a synthetic but representative decision trace to demonstrate coverage by the ten required fields and six groups, plus a new 'Limitations' subsection noting the value of future empirical audits on proprietary logs. This provides concrete illustration while remaining honest about the scope of validation performed. revision: partial
Circularity Check
No circularity; design proposal with external evaluation benchmark.
full rationale
The paper follows a design science methodology to define the Fragmented Trace Problem and propose the Decision Event Schema (DES) as a JSON Schema that covers four explicitly listed infrastructure layers within a single event structure. The uniqueness claim rests on an evaluation against 25+ existing formats rather than any self-referential equation, fitted parameter renamed as prediction, or self-citation chain. No load-bearing step reduces the result to its own inputs by construction; the schema definition, degradation-aware design, and tiered evidence strategy are presented as deliberate design choices, not derived predictions. The derivation chain is self-contained against the stated external benchmark.
Axiom & Free-Parameter Ledger
axioms (2)
- domain assumption The four infrastructure layers comprehensively capture governance-relevant decision information.
- standard math JSON Schema is a suitable and sufficient format for a governance logging standard.
invented entities (3)
-
Decision Event Schema (DES)
no independent evidence
-
Degradation-aware field design
no independent evidence
-
Tiered evidence strategy (lightweight, sampled, full)
no independent evidence
Forward citations
Cited by 2 Pith papers
-
Label-Free Detection of Governance Evidence Degradation in Risk Decision Systems
A composite multi-proxy framework detects harmful drift in label-free risk decision systems and enables graduated governance alerts.
-
Governed Auditable Decisioning Under Uncertainty: Synthesis and Agentic Extension
Synthesizes a governance evidence framework revealing a coverage gradient from full auditability in rule engines to structural breaks in agentic AI, with a cascade of uncertainty and four formal propositions.
Reference graph
Works this paper leans on
- [1]
-
[2]
Ahmad, A., Saad, M., & Mohaisen, A. (2019). Secure and Transparent Audit Logs with BlockAudit. Journal of Network and Computer Applications , 145, 102406–102406. https: //doi.org/10.1016/J.JNCA.2019.102406
-
[3]
Alu, F.F., & Oluwadare, S. (2026). An auditable and source-verified framework for clinical AI decision support. Frontiers in Artificial Intelligence , 9, 1737532–1737532. https://doi.or g/10.3389/frai.2026.1737532
-
[4]
Bisht, H. (2026). Governance-By-Design For AI-Based Insurance Fraud Detection: Auditabil- ity, Accountability, And Regulatory Traceability. Journal of International Crisis and Risk Communication Research, 214–222. https://doi.org/10.63278/jicrcr.vi.3620
- [5]
-
[6]
Car, N.J., Stenson, B., & Mirza, F. (2017). Modelling causes for actions with the Deci- sion and PROV ontologies. MODSIM2017, 22nd International Congress on Modelling and Simulation. https://doi.org/10.36334/modsim.2017.c2.car CNCF OpenTelemetry Project (2019). OpenTelemetry Specification. https://opentelemetr y.io/docs/specs/otel/ CNCF Serverless Working...
-
[7]
Devin, A. (2026). Epistemic Debt: The Economics of Ungoverned AI. Social Science Research Network. https://doi.org/10.2139/ssrn.6135728 European Parliament and Council of the European Union (2024). Regulation (EU) 2024/1689 of the European Parliament and of the Council of 13 June 2024 laying down harmonised rules on artificial intelligence and amending Re...
-
[8]
Fatmi, A. (2026). Faramesh: A Protocol-Agnostic Execution Control Plane for Autonomous Agent Systems. arXiv preprint (2601.17744) . https://doi.org/10.48550/arXiv.2601.17744
-
[9]
Gyevnar, B., Ferguson, N., & Schafer, B. (2023). Bridging the Transparency Gap: What Can Explainable AI Learn from the AI Act?. Frontiers in artificial intelligence and applications , 965–971. https://doi.org/10.3233/faia230367
-
[10]
Hartmann, D., Pereira, J.R.L.D., Streitbörger, C., & Berendt, B. (2024). Addressing the regulatory gap: moving towards an EU AI audit ecosystem beyond the AI Act by including civil society. AI and Ethics , 5, 3617–3638. https://doi.org/10.1007/s43681-024-00595-3
-
[11]
Hevner, A.R., March, S.T., Park, J., & Ram, S. (2004). Design Science in Information Systems Research. MIS Quarterly . https://doi.org/10.2307/25148625
-
[12]
Huynh, T.D., Tsakalakis, N., & Helal, A. (2020). Addressing Regulatory Requirements on Explanations for Automated Decisions with Provenance. Digital Government: Research and Practice, 2, 1–14. https://doi.org/10.1145/3436897 IEEE (2021). IEEE Standard for Transparency of Autonomous Systems. https://doi.org/ 10.1109/IEEESTD.2022.9726144 Joint Task Force (2...
-
[13]
Joseph, J. (2023). Trust, but Verify: Audit-ready logging for clinical AI. World Journal of Advanced Engineering Technology and Sciences , 10, 449–474. https://doi.org/10.30574/wja ets.2023.10.2.0249
-
[14]
Koisser, D., & Sadeghi, A.-R. (2023). Accountability of Things: Large-Scale Tamper-Evident Logging for Smart Devices. arXiv preprint (2308.05557) . https://doi.org/10.48550/arXiv.2 308.05557
-
[15]
Kulothungan, V. (2023). Using Blockchain Ledgers to Record AI Decisions in IoT. MDPI IoT, 6, 37–37. https://doi.org/10.3390/iot6030037
-
[16]
Malhotra, R.M. (2025). SHIT Theory: Systems Hurt In Theory: A Comprehensive Frame- work for Understanding Cybersecurity Governance Failure. Available at SSRN 5978876 , 1–25. https://doi.org/10.2139/ssrn.5978876
-
[17]
Moreau, L., & Missier, P. (2013). PROV-DM: The PROV Data Model. https://www.w3.o rg/TR/prov-dm/ Mökander, J., Axente, M., Casolari, F., & Floridi, L. (2021a). Conformity Assessments and Post-market Monitoring: A Guide to the Role of Auditing in the Proposed European AI Regulation. Minds and Machines , 32, 241–268. https://doi.org/10.1007/s11023-021-09577-...
-
[18]
Paccagnella, R., Datta, P., Hassan, W.U., Bates, A., Fletcher, C.W., Miller, A., & Tian, D. (2020). Custos: Practical Tamper-Evident Auditing of Operating Systems Using Trusted Execution. Network and Distributed System Security Symposium (NDSS) . https://doi.org/ 10.14722/ndss.2020.24065
-
[19]
Pratti, L., Bagchi, S., & Latif, Y. (2025). Data and Decision Traceability for SDA TAP Lab’s Prototype Battle Management System. arXiv. https://doi.org/10.48550/ARXIV.250 2.09827
-
[20]
Putz, B., Menges, F., & Pernul, G. (2019). A secure and auditable logging infrastructure based on a permissioned blockchain. Computers & Security , 87, 101602–101602. https: //doi.org/10.1016/j.cose.2019.101602
-
[21]
Schneier, B., & Kelsey, J. (1999). Secure Audit Logs to Support Computer Forensics. ACM Transactions on Information and System Security , 2, 159–176. https://doi.org/10.1145/31 7087.317089
work page doi:10.1145/31 1999
-
[22]
Sakata, T. (2010). Dapper, a Large-Scale Distributed Systems Tracing Infrastructure. Google Technical Report. https://doi.org/10.1145/2335356.2335365
-
[23]
Solozobov, O. (2026a). Decision Event Schema. GitHub. https://doi.org/10.5281/zenodo.1 8923178
-
[24]
Sutton, A., & Samavi, R. (2018). Tamper-Proof Privacy Auditing for Artificial Intelligence Systems. International Joint Conference on Artificial Intelligence (IJCAI), 5374–5378. https: //doi.org/10.24963/ijcai.2018/756
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.