SV-10 · Version 1.0 · 2026

The AI Agent
Accountability Standard

Published by SealVera · February 2026 · CC BY 4.0

Ten requirements every AI agent system must meet before making consequential decisions in production. Free to use, cite, and implement.

Self-assess your system →

Why this standard exists. AI agents are making consequential decisions — approving loans, authorizing medical procedures, screening job candidates, flagging transactions. Regulators, courts, and customers are increasingly demanding accountability for these decisions. Yet no widely-accepted standard exists for what "accountable" means in practice.

The SV-10 defines the minimum requirements for a production AI agent system to be considered accountable. It is not a certification. It is a checklist. It is designed to be used by compliance officers assessing readiness, engineering teams building systems, and executives making deployment decisions.

How to use it. Review each requirement. Assess whether your current system meets it. Where it does not, you have a gap. Gaps are not disqualifying — they are a roadmap. The goal is not perfection at launch. The goal is knowing where you stand before someone else finds out.

This standard is published under Creative Commons CC BY 4.0. You may use it, reproduce it, share it, and build on it — with attribution.

Section 1 — Decision Records
AA-01
Every consequential AI decision must produce a complete decision record
A decision record must capture, at minimum: the agent identity, the timestamp, the complete input data, the output and outcome, and the model or system version used. Records must be created automatically at the time of decision — not reconstructed afterward.
Rationale
Reconstructing a decision after the fact requires access to system state, model versions, and data that may no longer be available. Records created at decision time are the only reliable source of truth. "We can reproduce it" is not the same as "we recorded it."
Records Traceability
EU AI Act Art. 12 · FINRA Rule 4511
AA-02
Decision records must include structured, factor-level reasoning tied to actual input values
A record that states only the outcome ("REJECTED") without the reasoning is insufficient for accountability purposes. Each factor that influenced the decision must be documented with the actual value observed — not a general description. The reasoning must be traceable to specific data points in the input record.
Rationale
GDPR Article 22 and emerging AI regulations require "meaningful information about the logic involved." A confidence score or a summary sentence does not meet this bar. Factor-level reasoning tied to actual values is the only form of explanation that can be independently verified against the input data.
Explainability Evidence
GDPR Art. 22 · EU AI Act Art. 13
AA-03
Decision records must be tamper-evident — any modification must be cryptographically detectable
Every decision record must be signed or hashed at the time of creation such that any subsequent modification — to any field — is detectable by a third party. The signing mechanism must use an industry-standard algorithm. The verification key must be independently accessible. Records stored in mutable systems without cryptographic protection do not meet this requirement.
Rationale
A log that can be modified is not an audit trail — it is a file. Legal defensibility requires proof that a record represents what the system actually produced, not a edited version of it. Cryptographic signatures are the only mechanism that enables independent third-party verification.
Integrity Cryptographic
SOC 2 CC7.2 · FINRA Rule 17a-4
AA-04
The completeness of the decision record set must be verifiable — deletions must be detectable
It is not sufficient to prove that individual records are unmodified. The system must also provide a mechanism to detect whether any records have been deleted. This requires a hash chain, sequence counter, or equivalent mechanism that makes gaps in the record set detectable.
Rationale
Selective deletion of unfavorable decisions is a form of evidence tampering. Protecting individual records from modification while allowing deletion of inconvenient ones provides false assurance. Completeness verification closes this gap.
Completeness Chain Integrity
SOC 2 CC7.2 · EU AI Act Art. 12
Section 2 — Retention
AA-05
Decision records must be retained for the full duration required by applicable regulation
For high-risk AI systems as defined by the EU AI Act, decision records must be retained for a minimum of 10 years. For financial services AI systems, applicable FINRA or SEC retention periods apply. For systems processing personal data, GDPR retention limitation principles apply. Organizations must document the applicable retention period and demonstrate that their current record coverage meets it.
Rationale
EU AI Act Article 12 makes 10-year retention a legal requirement for high-risk systems. Many teams begin logging only after a compliance question arises — at which point years of records are permanently unavailable. The retention clock starts at first deployment, not at first audit.
Retention Regulatory
EU AI Act Art. 12 · Enforcement: Aug 2026
Section 3 — Behavioral Monitoring
AA-06
Agent behavior must be monitored against a documented baseline
Every production AI agent must have a documented behavioral baseline — the expected distribution of outcomes, approval rates, confidence levels, and decision volume under normal operating conditions. Deviations from the baseline must trigger investigation. The baseline must be updated when intentional system changes are made, and the history of baseline changes must be retained.
Rationale
Model providers update models without notice. Prompt drift occurs gradually. Data distributions shift. Without a documented baseline, organizations cannot demonstrate that their AI is operating as designed — a requirement for regulatory approval in multiple jurisdictions.
Monitoring Behavioral
EU AI Act Art. 9 · SOC 2 CC7.1
AA-07
Anomalies must be detected and alerted on before external parties report them
The organization must maintain an active monitoring system that detects behavioral anomalies — unusual decision rates, outcome distribution shifts, confidence degradation, agent silence — and generates alerts to responsible parties. Finding out about a system anomaly from a customer complaint, news story, or regulatory inquiry is not acceptable. Alert history must be logged with timestamps and acknowledgment records.
Rationale
Demonstrating due diligence requires showing that anomalies were caught and addressed internally. An alert history is evidence of active oversight — the kind of evidence that distinguishes a negligent operator from a responsible one in regulatory and legal proceedings.
Alerts Oversight
EU AI Act Art. 9 · SOC 2 CC7.3
Section 4 — Traceability
AA-08
Multi-agent workflows must be traceable as a single decision chain
When multiple AI agents contribute to a single decision — fraud screening, risk scoring, final approval — the full chain of agent decisions must be traceable as a unified sequence. It must be possible to reconstruct which agent made which decision, in what order, with what inputs, at what time. Isolated records from individual agents that cannot be linked to a shared workflow do not meet this requirement.
Rationale
Multi-agent architectures are increasingly common in production. When something goes wrong in a pipeline, accountability requires identifying which agent in the chain made the consequential decision. Individual agent logs that cannot be correlated are insufficient for this purpose.
Tracing Workflow
EU AI Act Art. 13 · FINRA OATS
AA-09
Any past decision must be reproducible from its original inputs
The organization must be able to replay any recorded decision using the original input data to verify that the AI's reasoning is consistent and accurate. The original inputs must be preserved intact in the decision record. The replay mechanism must confirm whether the decision is consistent with the original outcome and flag discrepancies. This capability must be available for the full retention period.
Rationale
The ability to verify a past decision is distinct from the ability to record it. Replay enables investigation — when a decision is challenged, you can demonstrate what the system would produce given the same inputs, which is a powerful form of evidence in both regulatory and legal contexts.
Reproducibility Verification
GDPR Art. 22 · EU AI Act Art. 14
Section 5 — Reporting and Response
SV-10
Compliance reports must be generatable on demand, not assembled under pressure
The organization must be able to produce a complete audit report for any time window — covering all AI decisions made, the reasoning behind each, chain integrity verification, retention coverage, and behavioral monitoring status — within hours, not weeks. The report must be in a format acceptable to regulators and legal counsel. The engineering team must not be required to assemble it manually each time it is needed.
Rationale
Regulatory inquiries and legal discovery come with time pressure. An organization that requires weeks of engineering work to respond to an audit request is operationally non-compliant regardless of the quality of its underlying records. The ability to respond quickly is itself a compliance requirement under several frameworks.
Reporting Response
EU AI Act Art. 12 · SOC 2 CC4.1
Self-assessment tool

Where does your system stand?

Check each requirement your system currently meets. See your readiness score. Use the gaps as a roadmap.

SV-10 Readiness Score 0 / 10
Check the requirements above to see your score.
See how SealVera closes your gaps →
SV-10 COMPLIANT SealVera · 2026
SV-10 Compliant Badge
SV-10 AUDIT READY SealVera · 2026
Audit Ready Badge
SV-10 COMPLIANT AI Agent Accountability Standard · SealVera 2026
Horizontal Badge

SealVera implements the SV-10.

Every requirement in this standard maps directly to a capability in SealVera. Connect your agents and your compliance team has everything they need to meet all ten.

Start for free → See a live audit trail