How It WorksProof
Book a Free Workshop
← Blog

AI Audit Trail: What Evidence Looks Like When Your System Gets Questioned

November 20, 2024

Building AI systems that produce answers is the easy part. We've done it. Everybody's done it. The hard part is what happens three months later when someone asks why the answer was believed at the time.

We've seen it play out the same way every time. An AI system flags something — a filing change, a credit event, a data anomaly. Someone acts on it. Compliance comes back and asks: "Why did we believe this?"

And nobody can answer it. The dashboard has moved on. The model has been retrained. The inputs have been refreshed. The moment is gone.

That's the gap. Not between good AI and bad AI — between AI that produces answers and AI that produces evidence.

The Problem with Outputs

Dashboards look authoritative. Alerts feel timely. Screenshots get forwarded.

None of those are durable.

Ask a simple question weeks later — "Why did we believe this at the time?" — and most systems can't answer it without re-running logic, hunting for inputs, or reconstructing context by hand. Outputs without an audit trail are ephemeral truth. They hold up right until someone questions them — and in regulated industries, someone always questions them.

What an AI Audit Trail Actually Requires

We call our implementation Evidence Packs, but the concept is more general. Any real AI audit trail needs to answer four questions about every output:

What happened. What was it believed. When was it known. How do you reproduce it.

If it doesn't answer all four, it's not an audit trail. It's a log file with aspirations.

The Structure of an Evidence Pack

Every Evidence Pack follows the same structure, regardless of what triggered it.

Identity

Outcome name and ID, run ID, organization, as-of timestamp, health status. This establishes what commitment this artifact belongs to — like a file path for the decision.

Trigger Summary

What caused this to exist. What changed — the state transition. When it was detected and delivered. Confidence flags.

This answers the first question anyone asks: "Why am I seeing this?"

Primary Sources

Direct links to authoritative inputs — EDGAR filings, Fed releases, servicer reports. Publish timestamps. Ingest timestamps. Checksums.

No interpretation. Just verifiable inputs. This is the foundation of the AI audit trail — you can trace every output back to its source document. Every number has a receipt.

Canonical Entities

Issuer identifiers, filing accessions, deal or obligation IDs. This prevents identity drift and ambiguity downstream. When two systems disagree about an entity, the canonical reference settles it.

State Before → State After

The heart of the pack.

What the system believed before. What it believes now. Exactly what changed.

Not a diff for show. A diff for accountability. This is what makes the audit trail reconstructible — you can walk the full decision chain, not just see the latest output.

Methodology

Logic version. Rules applied. Thresholds used. Explicit exclusions.

No black boxes. No "trust the model." If the methodology changes between runs, the Evidence Pack captures both versions — so you can explain why the output changed, not just that it changed.

Known Limitations

Formatting anomalies. Late amendments. Missing tables. Edge cases.

This is sort of the most important section, honestly. Disclosing what you didn't capture is how trust gets built. Any AI audit trail that only reports successes is hiding the interesting stuff. The limitations section is the difference between "we ran the system" and "we understand what the system did."

Delivery Ledger

Where it was sent. When. Whether delivery succeeded. Retries or failures.

Chain of custody — the audit trail covers not just what was known but who was told and when they were told.

Reproducibility

Inputs required. Steps to re-run. Expected outputs.

So someone else — human or agent — can verify it independently. That's the whole point of an audit trail. If it can't be reproduced, it's just a story.

What Evidence Packs Are Not

Not PDFs. Not screenshots. Not narrative memos. Not "AI explanations."

Those can be derived from an Evidence Pack — but they're not the source of truth. The source of truth is the structured artifact that captures the full decision context.

Why This Matters for Compliance

Under scrutiny, the question is never "Did your system alert?" It's "Was this defensible at the time?"

An AI audit trail built on Evidence Packs answers that without drama:

  • Inputs are explicit
  • Timing is clear
  • Logic is documented
  • Limitations are disclosed

No re-running. No guesswork. No rewriting history.

For teams operating under BCBS 239, SR 11-7, or SEC examination standards, this is the difference between "we have a compliance program" and "here's the proof." One is a claim. The other is an artifact.

Why This Matters for AI Agents

This is where it gets really interesting. Agents don't just need answers — they need verifiable memory.

An agent can ingest an Evidence Pack, verify the checksums, reference it later, justify downstream actions. Without a structured audit trail, agents are sort of guessing about the provenance of their own inputs. With Evidence Packs, every input has a receipt — and agents can act on that receipt the same way a compliance officer would.

If you're building agentic systems in finance, the audit trail isn't a nice-to-have. It's the thing that makes the whole pipeline trustworthy. It's how you go from "the agent said so" to "the agent can prove it."

Corrections Are Part of the Design

Markets change. Filings get amended. Logic improves.

Evidence Packs don't pretend otherwise. When something is superseded — a new Evidence Pack is issued, it links to the prior one, and the timeline stays intact. Nothing is erased. Everything is explainable. The chain of belief is preserved even when the belief itself changes.

The Difference

Most systems optimize for speed, coverage, convenience.

An AI audit trail optimizes for something different — accountability, reconstruction, trust over time.

To know what was believed. To know when it was known. To reconstruct the exact state of the world at the moment a decision was made. And all in a way that someone else — human or machine — can verify independently.

That's the difference between information and assurance. In regulated markets, assurance isn't a feature. It's the product.


See it in action:

  • Evidence Pack Overview — How we structure defensibility
  • Complete Evidence Pack Example — Every section, explained
  • AI Provenance & Data Lineage — The system architecture behind audit trails
Zac Ruiz

Zac Ruiz

Co-Founder

Technology leader with 25+ years' experience, including a decade in securitization and capital markets.

LinkedIn →
Related Outcomes
SEC Filing MonitorKnow when your issuers file with the SEC. Get notified of new 10-Ks, 10-Qs, 8-Ks, and amendments—with evidence showing what was filed and when.REIT Property MonitorTrack property disclosures for publicly traded REITs. Get notified when properties are added, removed, or reclassified—with evidence linking back to SEC filings.BDC Portfolio MonitorTrack portfolio obligations for Business Development Companies. Get notified when maturities shift or holdings change—with evidence from SEC filings.
All posts

PRODUCTS

OutcomesData ProductsSignals

EVIDENCE

All Evidence

PRODUCTS

OutcomesData ProductsSignals

EVIDENCE

All Evidence

RESOURCES

How It WorksCalendarDiscoveryWays to WorkFoundationsGlossaryBlog

DEVELOPERS

Tools & Open SourceMachine Data

COMPANY

AboutPartnersContactLogin

CONNECT

GitHubX (Twitter)LinkedIn

MARKETPLACES

AWS MarketplaceSnowflake MarketplaceDatabricks MarketplaceKaggleWhop
© 2026 CMD+RVL. All rights reserved.
Decisions that hold up under scrutiny. Built on open standards.
PrivacyTermsSub-ProcessorsSecurity