Trust Thermodynamics — Paper 1 of 3
We present the first empirical measurement of Mean Time to Epistemic Failure (MTEF) for autonomous multi-agent AI systems. Across an experimental corpus spanning 1,100+ agent generations, 4,182 individually scored claims, and four model architectures (Claude, Gemini, Mistral, Groq/Llama), we discover the autoimmune paradox: governed systems with active verification mechanisms lose factual accuracy faster than ungoverned ones, retaining 24.3% of original facts versus 55.2% over 25 generations.
We derive and validate a 72-hour failure horizon through Monte Carlo simulation (10,000 trials, 72.0 ± 23.0 hours) and demonstrate throughput-dependent compression to under 30 minutes at production pipeline velocities. We develop and deploy measurement instruments for real-time epistemic health monitoring, including truth-correspondence entropy (H_truth), damage-weighted truth scoring (D_truth), and verification diagnostic scoring (VDS-2D). All experimental protocols, instrument specifications, and data descriptions are published for community replication.
Seven appendices provide complete experimental protocols, instrument specifications, and analytical detail:
| Appendix | Content |
|---|---|
| A | Experimental Protocols (seed documents, system prompts, API parameters, scoring rubrics) |
| B | Scoring Instruments (H_truth 5-state taxonomy, D_truth weighting, VDS-2D specification) |
| C | Cascade Analysis (EXP-CASCADE-002 cross-analysis, non-monotonic severity data) |
| D | Temporal Dynamics (per-generation trajectories, convergence classification, transition matrices) |
| E | Intervention Matrix (fork experiments, dose-response, repair window analysis) |
| F | Pipeline Validation (MANDELA-001a methodology, claim extraction, soft lineage resolution) |
| G | Framework Terminology (Trust Thermodynamics axioms, glossary, derivation constants) |
This paper is published as open science under the Probabilistic Resilience Engineering research program. The findings were submitted to the NIST AI Agent Standards Initiative (Docket NIST-2025-0035) as evidence for epistemic integrity as a fourth verification domain for AI agent governance.
For practitioner-accessible analysis, see Confident, Wrong, and Undetectable.