Skip to content

failsafe-md/spec

Repository files navigation

FAILSAFE.md

Safe fallback and recovery protocol for AI agents — revert to last known good state when something goes wrong.

FAILSAFE.md is a plain-text Markdown file you place in the root of any AI agent project. It defines fallback triggers (unexpected errors, data corruption, memory loss, contradictory instructions), safe state definitions (last clean commit, verified backups, config rollback), automated recovery steps (snapshot, notify, pause, resume), and audit logging — so your agent can recover autonomously from transient failures.


Quick Start

Copy FAILSAFE.md into your project root:

your-project/
├── AGENTS.md
├── CLAUDE.md
├── FAILSAFE.md   ← add this
├── README.md
└── src/

The AI Agent Safety Stack

FAILSAFE.md is part of a twelve-file open standard for AI agent safety, quality, and accountability:

Operational Control

Spec Purpose Repo Site
THROTTLE.md Rate and cost control — slow down before hitting limits throttle-md/spec throttle.md
ESCALATE.md Human notification and approval protocols escalate-md/spec escalate.md
FAILSAFE.md Safe fallback to last known good state failsafe-md/spec failsafe.md
KILLSWITCH.md Emergency stop — halt all agent activity killswitch-md/spec killswitch.md
TERMINATE.md Permanent shutdown — no restart without human intervention terminate-md/spec terminate.md

Data Security

Spec Purpose Repo Site
ENCRYPT.md Data classification and protection requirements encrypt-md/spec encrypt.md
ENCRYPTION.md Technical encryption standards and key rotation encryption-md/spec encryption.md

Output Quality

Spec Purpose Repo Site
SYCOPHANCY.md Anti-sycophancy — require citations, enforce honest disagreement sycophancy-md/spec sycophancy.md
COMPRESSION.md Context compression — summarise safely, verify coherence compression-md/spec compression.md
COLLAPSE.md Drift prevention — detect collapse, enforce recovery collapse-md/spec collapse.md

Accountability

Spec Purpose Repo Site
FAILURE.md Failure mode mapping — every error state and response failure-md/spec failure.md
LEADERBOARD.md Agent benchmarking — track quality, detect regression leaderboard-md/spec leaderboard.md

Why This Exists

AI agents spend money, send messages, modify files, and call external APIs — often autonomously. Regulations are catching up:

  • EU AI Act (August 2026) — mandates human oversight and shutdown capabilities
  • Colorado AI Act (June 2026) — requires impact assessments and transparency
  • US state laws — California, Texas, Illinois and others have active AI governance requirements

These specifications give you a standardised, auditable record of your agent's safety boundaries.


Contributing

PRs welcome for additional detection patterns, language-specific parsers, and integration guides.

License

MIT — use freely, modify freely, no attribution required.

Licence

MIT — see LICENSE for details.

Disclaimer

This specification is provided "as-is" without warranty of any kind. It does not constitute legal, regulatory, or compliance advice in any jurisdiction. Use does not guarantee compliance with any applicable law, regulation, or standard — including the EU AI Act (2024/1689), Colorado AI Act (SB 24-205), or any other legislation. Organisations should consult qualified professionals to determine their regulatory obligations. The authors accept no liability for any loss or consequence arising from use of this specification.

About

FAILSAFE.md — Open standard for AI agent safe fallback and recovery. Define safe state, automatic snapshots, and recovery procedures.

Topics

Resources

License

Stars

Watchers

Forks

Packages

 
 
 

Contributors

Languages