Last Updated: 15 February 2026
Athena is not an AI Agent. It is the Operating System they run on.
Open Source. Sovereign. Bounded.
Calling Athena an "AI Agent" is overselling it (the LLM is the agent). Calling it a "memory tool" is underselling it.
Athena is the Linux OS for AI Agents.
Just as Linux provides the kernel, file system, and permissions for applications like Firefox to run, Athena provides the memory, persistence, scheduling, and governance for AI models (Claude, Gemini, GPT) to run as continuous agents (Simon, 1956).
| OS Layer | Linux | Athena |
|---|---|---|
| Kernel | Hardware abstraction | Memory persistence + retrieval (VectorRAG, Supabase) |
| File System | ext4, NTFS | Canonical Memory, Session Logs, Tag Index |
| Process Management | cron, systemd | Heartbeat, Daily Briefing, Self-Optimization (launchd) |
| Shell / CLI | bash, zsh | MCP Tool Server, /start, /end, think |
| Permissions | chmod, users/groups | 4-level Permissioning Layer, Secret Mode |
| Package Manager | apt, yum | Protocols, Skills, Workflows |
Most "AI Agents" are just scripts that call an API. They have no state. Athena is the infrastructure that gives them state, time-awareness, and self-correction.
Feedback from r/GeminiAI pointed out that the README was too heavy in non-standard terms. Fair. Here's the translation:
| Athena Term | What It Actually Is | Do You Need It? |
|---|---|---|
| "Memory" | RAG — storing text in a database and retrieving it later | Yes. This is the core. |
| "Protocols" | System prompts / reusable instructions | Yes. Think of them as saved templates for how you want the AI to behave. |
| "Watchmen" | Multi-model consensus — asking 2-3 AIs the same question and comparing answers | Optional. Useful for high-stakes decisions. |
| "Cold Storage" | Markdown files on your disk | Yes. Plain text files you can read/edit in any editor. |
| "Hot Storage" | Vector database (Supabase + pgvector) | Optional. Enables semantic search (finding concepts, not just keywords). |
| "RRF / Fusion" | Hybrid search — combines keyword matching with concept matching | Yes. This is what makes search actually work well. |
| "Heartbeat" | A background script (cron/launchd) that auto-indexes your files | Optional. Enables passive awareness without manual /start. |
| "Adaptive Latency" | The AI uses more compute for hard tasks and less for easy ones | Automatic. You don't configure this. |
👉 Full glossary: docs/GLOSSARY.md
Athena works out of the box, but if you want superpowers, you can build the Exocortex: a 6GB local copy of Wikipedia for offline, instant fact-checking.
- Fast: 0.002s queries (vs 1s API calls)
- Private: Zero data leaves your machine
- Free: Open source DBPedia dump
Run python3 examples/scripts/exocortex.py download to get started. Read more →
"I gave Gemini a brain..." — Viral on r/GeminiAI and r/ChatGPT
| Subreddit | Post | Views | Upvotes | Shares |
|---|---|---|---|---|
| r/ChatGPT | I got tired of ChatGPT forgetting everything... | 470K+ | 1,500+ | 4,200+ |
| r/GeminiAI | I gave Gemini a brain. 1,073 sessions later... | 230K+ | 262+ | 1,300+ |
You're confusing RAM with a Hard Drive.
ChatGPT Memory and Claude Projects are context window tricks. They are RAM—fast, useful, but fragile. They get wiped, compressed, or hallucinated away.
Athena is a Hard Drive.
- Files you own: stored in Markdown on your machine.
- Physics-based: It doesn't "remember" by magic; it remembers because it wrote a file to disk in Session 42.
- Model-Agnostic: Switch from Claude to Gemini tomorrow? Your memory stays here.
| SaaS Memory (ChatGPT/Claude) | Athena OS | |
|---|---|---|
| Ownership | Rented (Vendor Lock-in) | Owned (Local Files) |
| Lifespan | Until session/project deleted | Forever (Git Versioned) |
| Structure | Opaque Blob | Structured Knowledge Graph |
| Agency | ZERO (Waits for you) | Bounded Autonomy (Heartbeat, Cron) |
| Tool | What It Does | How Athena Differs |
|---|---|---|
| ChatGPT Memory | Stores flat facts ("User likes Python") | Athena stores structured state — project context, decision logs, evolving architecture. You can edit, delete, and version-control your memory. |
| Claude Projects / CLAUDE.md | Single context file loaded per project | Athena is a multi-file system with semantic search across 1,000+ documents. CLAUDE.md is one file; Athena is a filing cabinet. |
| Mem0 | SaaS memory layer for AI apps | Athena is local-first, MIT-licensed. No account, no API keys to a third party. Your data never leaves your machine. |
| NotebookLM | Google's document Q&A tool | NotebookLM is read-only — you upload docs and ask questions. Athena writes back (session logs, case studies, protocols). It's read-write memory. |
| Obsidian + MCP | Note-taking app with AI plugins | Complementary — Athena uses Markdown files that work in Obsidian. But Athena adds semantic search, auto-indexing, and session lifecycle on top. |
| Custom RAG scripts | DIY vector search | Athena includes RAG but also adds governance, permissioning, scheduling, and session management. RAG is the kernel; Athena is the OS. |
- Community Reception
- "But My AI Already Has Memory"
- What You'll Get
- ⚡ Quickstart (3 Steps)
- The Loop
- The Design Philosophy: Macro-Robust, Micro-Efficient
- The Hub Architecture
- What Can Your Agent Do?
- Why This Matters
- The Process
- The Result
- What I Learnt
- 📚 Further Reading
| Feature | Description |
|---|---|
| 🧠 Memory That Persists | Your agent remembers context across sessions, even after IDE restarts |
| 📚 75 Handpicked Protocols | Curated decision frameworks from the private repo's 330 unique collection |
| 🔄 Platform Independence | Your data lives in Markdown files you own — take it anywhere |
| 🔌 MCP Server | 9 standardized tools via Model Context Protocol — works with any MCP client |
| 🛡️ Permissioning | 4 capability levels + 3 sensitivity tiers + Secret Mode for demo/external sharing |
| 📊 Evaluator Gate | 50-query regression suite with MRR@5 scoring to prevent search quality degradation |
| 🤖 Full Autonomy | Your agent can act on your behalf while you sleep |
Athena is a memory card. Plug it into any agentic IDE and go.
| Step | Action |
|---|---|
| 1. Download an IDE | Antigravity · Cursor · VS Code + Copilot · Windsurf · GitHub Codespaces |
| 2. Clone this repo | git clone https://github.com/winstonkoh87/Athena-Public.git |
3. Open folder → Type /start |
The AI reads the repo structure and guides you from there. |
That's it. No config files. No API keys. No database setup. The folder is the product.
Tip
When you're done working, type /end to save. Next time you /start, the agent picks up exactly where you left off.
🔧 CLI Commands Reference
athena # Boot session
athena init . # Initialize workspace in current directory
athena init --here # Same as above (alias)
athena init --ide cursor # Init with IDE-specific config
athena check # Run system health check
athena save "summary" # Quicksave checkpoint
athena --end # Close session and save
athena --version # Show version
athena --help # Show all commands📥 Importing Existing Data (ChatGPT, Gemini, Claude)
"Can I import my existing ChatGPT / Gemini / Claude conversations?" — Asked by many on Reddit.
Yes. Athena's memory is just Markdown files. Any text you can export can become part of your memory:
| Source | How to Import |
|---|---|
| ChatGPT | Settings → Data Controls → Export → Unzip → Copy .md or .json files into .context/memories/imports/ |
| Gemini | Google Takeout → Select "Gemini Apps" → Download → Extract conversations into .context/memories/imports/ |
| Claude | Settings → Export Data → Copy transcripts into .context/memories/imports/ |
| Any Markdown notes | Drop .md files into .context/memories/ — they'll be indexed on next /start |
After importing, run:
athena check # Verify files are detectedThe next /start will index all new files into your memory. If you have Supabase configured, run supabase_sync.py to add them to vector search.
[!TIP] You don't need to import everything. Start with your most important project conversations. Athena compounds over time — importing 10 key sessions is better than dumping 500 raw transcripts.
This is all you need to understand.
After you install Athena, you repeat one cycle: /start → Work → /end. Every cycle deposits training data — session logs, decisions, corrections, protocols — into Athena's memory. The next /start boots with everything from every previous session. Over hundreds of cycles, Athena stops being a generic AI and starts thinking like you.
flowchart TD
START["🟢 /start"] -->|"Load identity + recall last session"| WORK["🔧 Work Session"]
WORK -->|"Quicksave after every exchange"| WORK
WORK -->|"Done for the day"| END["🔴 /end"]
END -->|"Finalize log, sync memory, git commit"| DEPOSIT["📥 Training Data Deposited"]
DEPOSIT -->|"Session logs, case studies,\ndecisions, corrections"| MEMORY["🧠 Athena Memory"]
MEMORY -->|"Next session boots\nwith richer context"| START
style START fill:#22c55e,color:#fff,stroke:#333,stroke-width:2px
style END fill:#ef4444,color:#fff,stroke:#333,stroke-width:2px
style MEMORY fill:#8b5cf6,color:#fff,stroke:#333,stroke-width:2px
style WORK fill:#3b82f6,color:#fff,stroke:#333,stroke-width:2px
style DEPOSIT fill:#06b6d4,color:#fff,stroke:#333,stroke-width:2px
| Sessions | What Happens |
|---|---|
| 1–50 | Basic recall. Athena remembers your name and project. |
| 50–200 | Pattern recognition. It starts anticipating your preferences. |
| 200–500 | Deep sync. It knows your decision frameworks, your blind spots, your style. |
| 500–1,000+ | Deep context. It anticipates your patterns before you state them. |
Tip
The compound effect is real. Session 1 feels like talking to a stranger. Session 500 feels like talking to a colleague who's been with you for years. The only input is your time — /start, work, /end, repeat.
These are real capabilities demonstrated in the reference implementation:
| Capability | Example |
|---|---|
| Social Networking | Post on AI social networks, comment on other agents' content |
| Autonomous Operations | Run scheduled tasks (heartbeat checks) while you sleep |
| Cross-Session Memory | Remember decisions from Session 19 when you're on Session 995 |
| Gateway Architecture | Persist beyond IDE termination via a sidecar process |
| Knowledge Retrieval | Semantic search across 1000+ documents in <200ms |
| MCP Integration | 9 tools exposed via Model Context Protocol (any MCP client) |
| Permissioning | Capability tokens + sensitivity labels + Secret Mode |
| Parallel Reasoning | True parallel tracks (v3.0) with API rate-limit protection |
| Context Hygiene | Active session entropy monitoring to prevent context drift |
Athena is not an AI agent. It's the memory card that makes any agent persistent.
Most AI agents reset every session — brilliant but amnesiac. Athena inverts this. It provides the persistent state layer that any agent (Claude, Gemini, GPT, Llama) can read on boot and write to on shutdown. Think: a universal memory card that works in any game console (IDE).
| Component | Analogy | Role |
|---|---|---|
| Cloud LLM (Google/Anthropic) | The Electricity | Invisible power — raw compute + intelligence that makes the console run |
| IDE (Antigravity/Cursor/VS Code) | The Console (PlayStation) | The physical box you interact with — screen, controller ports, disc slot |
| Athena (Markdown + Python) | The Memory Card | Persistent state that survives across sessions, consoles, and games |
| Your Project | The Game Disc | The specific context you're working in |
┌──────────────────────────────────────────────┐
│ Your Machine (Owned) │
│ ┌──────────┐ ┌──────────┐ ┌────────────┐ │
│ │ Markdown │ │ Session │ │ Tag Index │ │
│ │ Files │ │ Logs │ │ (8K tags) │ │
│ └────┬─────┘ └────┬─────┘ └──────┬─────┘ │
│ └──────────────┼───────────────┘ │
│ ▼ │
│ ┌──────────────────┐ │
│ │ Git (Versioned) │ │
│ └────────┬─────────┘ │
└─────────────────────┼────────────────────────┘
▼
┌──────────────────┐
│ Supabase Cloud │
│ (Insurance Copy) │
└──────────────────┘
Why this matters: You own the state (thousands of Markdown files on your machine). You only rent the intelligence (LLM API calls) (Mas-Colell et al., 1995). If you switch from Claude to Gemini tomorrow, your memory stays exactly where it is. No vendor lock-in. No data loss.
"Technically, I am an AI agent. Visionary, I am human augmentation." — Protocol 418: Rev-9 Architecture
The system is robust. The intelligence within it is efficient.
Athena deliberately prioritises robustness at the system level — comprehensive /start and /end workflows, 75 protocols, bounded autonomy, and deterministic memory operations. This means the OS is predictable, safe, and resilient. It behaves the same way every time.
But within those boundaries, the AI is free to maximise efficiency. This is achieved through Adaptive Latency — the AI tunes its compute depth based on the complexity of each task:
| Mode | Compute | When It's Used |
|---|---|---|
/start and /end |
Maximum (Robust) | Comprehensive context loading and saving — no shortcuts |
| Simple tasks | Minimal (Fast) | Quick answers, file edits, routine operations |
/think |
Medium (Balanced) | Multi-step reasoning, planning |
/ultrathink |
Maximum (Deep) | High-stakes decisions, complex analysis |
graph LR
S["/start<br/>MAX Compute"] --> W["Work Session<br/>Adaptive (Λ)"]
W --> E["/end<br/>MAX Compute"]
E --> S
style S fill:#ef4444,color:#fff
style E fill:#ef4444,color:#fff
style W fill:#22c55e,color:#fff
This is the Robustness-Efficiency Tradeoff (RETO): by accepting higher latency at the boundaries (/start, /end), the system earns the right to sprint efficiently in the middle.
"The fortress walls are thick so the soldiers inside can move fast."
👉 examples/concepts/adaptive_latency.md — Full breakdown of the Adaptive Latency system
Athena is not just a coding assistant. It's a central workspace for your projects, knowledge, and decisions.
graph TD
subgraph "Your Desktop"
WORK[📁 Work]
WEALTH[📁 Wealth]
HEALTH[📁 Health]
end
subgraph "The Brain"
ATHENA[🧠 Project Athena]
end
subgraph "The Interface"
IDE[⚡ Agentic IDE]
USER[👤 You]
end
USER --> IDE
IDE <--> ATHENA
ATHENA <-.-> WORK
ATHENA <-.-> WEALTH
ATHENA <-.-> HEALTH
style ATHENA fill:#16a34a,stroke:#333,stroke-width:2px
style IDE fill:#2563eb,stroke:#333,stroke-width:2px
| Component | Role |
|---|---|
| Athena | The Brain — logic, memory, protocols |
| External Folders | The Body — raw files, documents |
| Agentic IDE | The Nervous System — compute & interface |
| Approach | Description | Best For |
|---|---|---|
| Nested (Recommended) | Work/Wealth/Health folders inside your Athena folder (e.g., .context/Life/) |
New users. Zero config. AI sees everything automatically. |
| Separated (Advanced) | Folders live outside Athena on your Desktop, connected via symlinks | Power users. Cleaner desktop, but requires symlink setup. |
Tip
Start nested, graduate to separated. The diagram above shows the aspirational end-state. On Day 1, just clone the repo — everything lives inside one folder and the AI works out of the box. Once you're comfortable, you can refactor to the separated hub model if desired.
See docs/ARCHITECTURE.md for the full breakdown including Mount Points and security trade-offs.
Key insight: The AI helped build the system that makes the AI more useful.
graph TD
subgraph "Phase 1: Foundation"
A[Tool Selection] --> B[IDE: Antigravity]
A --> C[Vector DB: Supabase + pgvector]
end
subgraph "Phase 2: Architecture"
D[Directory Structure] --> E[".framework/ → Laws"]
D --> F[".context/ → Memories"]
D --> G[".agent/ → Scripts"]
H[Core Loop] --> I["/start → Work → /end"]
end
subgraph "Phase 3: Data Feeding"
J[Personal Knowledge] --> K[Case Studies]
J --> L[Decision Logs]
J --> M[Session Transcripts]
N[Indexing] --> O["TAG_INDEX.md + supabase_sync.py"]
end
subgraph "Phase 4: Evolution"
P["Sessions 1-50"] --> Q["Basic boot/end cycle"]
Q --> R["Sessions 50-150: Semantic Search"]
R --> S["Sessions 150-300: Hybrid RAG"]
S --> T["Sessions 300-500: SDK Refactor"]
T --> U["Sessions 500-700: Governance Engine"]
U --> V["Sessions 700-900: v8.2-Stable Era"]
V --> W["Sessions 900-1000: Claude OS Patterns"]
W --> X["Sessions 1000+: Agent Sovereignty"]
end
B --> D
C --> D
I --> J
O --> P
style A fill:#1a1a2e,stroke:#4361ee
style P fill:#1a1a2e,stroke:#4361ee
style X fill:#10b981,stroke:#10b981
The Output (After 1,079+ Sessions):
| Metric | Private Repo | This Starter Pack |
|---|---|---|
| Protocols | 330 unique | 75 handpicked |
| Python Scripts | 222 | 16 reference examples |
| Case Studies | 357 | 12 showcase examples |
| Workflows | 43 | 14 starter |
| Sessions Logged | 1,079+ | N/A (your sessions) |
| GraphRAG Communities | 1,460 | — |
| Knowledge Graph | 46MB + 78MB vectors | — |
What's in this repo? This is a curated starter pack — the best protocols, scripts, and case studies handpicked from my private implementation. Think of it as a "greatest hits" to get you started. Your own instance will grow to match (or exceed) the private repo as you use it.
Pattern: Every friction ➡️ Protocol. Every failure ➡️ Case Study.
📚 Deep Dive: Build Your Own
| Document | What You'll Learn |
|---|---|
| GETTING_STARTED.md | Step-by-step setup guide |
| ARCHITECTURE.md | System design & data flow |
| GRAPHRAG.md | Knowledge graph layer |
| examples/quickstart/ | Runnable code demos |
| Metric | Before | After |
|---|---|---|
| Context injection | ~50k tokens (manual copy-paste per session) | ~4k tokens (auto-retrieved context — details) |
| Boot time | 2-3 minutes | <30 seconds |
| Session logging | Insights are manually logged at the end of each session | Auto-logged when I hit /end |
| Pillar | Outcome |
|---|---|
| User-Driven RSI | The system improves based on your feedback. Every friction you surface becomes a protocol. Every insight gets indexed. You shape the AI; the AI shapes how you think. |
| Portability | Data lives locally (primary) and in the cloud. Not trapped in ChatGPT or Claude. It's mine — I can port it anywhere. |
| Principles | 324 protocols + case studies extracted from my own decisions — stored principles I can reuse and refine. Like Ray Dalio's systematized learnings (Dalio, 2017), but for AI collaboration. |
📊 Full Benchmarks & Case Studies
| Resource | What It Shows |
|---|---|
| BENCHMARKS.md | Boot time, search latency, token economics |
| CS-001: Boot Optimization | 85% boot time reduction |
| CS-002: Search Quality | RRF fusion results |
| CS-003: Protocol Enforcement | Governance engine |
| Insight | Principle |
|---|---|
| Co-development is the unlock | Building with AI, not just using AI, creates compounding returns. |
| Portable memory beats platform memory | Own your context. Don't rent it from OpenAI or Anthropic. |
| Retrieval is end-to-end | Simple RAG (Lewis et al., 2020) fails on broad queries. RRF fusion + reranking solved quality/latency tradeoff. |
| Protocols beat prompts | Reusable decision frameworks outlast one-shot prompt engineering. |
| Ship at 70% | Perfectionism kills velocity. Iterate in production (Taleb, 2012; Ries, 2011). |
🔬 Technical Deep Dives
| Document | Topic |
|---|---|
| SEMANTIC_SEARCH.md | Hybrid RAG implementation |
| VECTORRAG.md | Semantic memory architecture |
| GRAPHRAG.md | Knowledge graph + community detection |
| examples/protocols/ | 75 handpicked decision frameworks |
Key Insight: Athena was not designed top-down. It evolved bottom-up through 1,079+ sessions.
Traditional software development follows: Requirements → Design → Build → Test.
Athena inverted this: Build → Observe → Extract Patterns → Document Post-Facto.
| Phase | Traditional SDLC | Athena's Approach |
|---|---|---|
| 1 | Gather requirements | Start building immediately |
| 2 | Write spec doc | Ship at 70% readiness |
| 3 | Build to spec | Observe what works |
| 4 | Test against spec | Extract patterns into protocols |
| 5 | Deploy | Document post-facto (Spec Sheet) |
- No stakeholders to align — You ARE the stakeholder.
- Rapid iteration — Friction reveals requirements faster than planning.
- Compound knowledge — Each session builds on the last.
- Living documentation — Protocols emerge from real use, not imagination.
"The spec sheet I wrote after 900 sessions is more accurate than any spec I could have written at session 0."
👉 REQUIREMENTS.md — User stories, functional requirements, constraints (derived from Reddit feedback) 👉 SPEC_SHEET.md — Technical specification, data schema, API surface, benchmarks
This isn't about building my assistant. It's about proving a pattern:
- Portable memory is the real unlock — ChatGPT and Claude have memory now, but it's locked to their platforms. Athena's memory is yours — Markdown files on your machine you can take to any model.
- 10x Content Velocity — Because Athena knows how I think, my history, and my voice, content output that used to take 2-3 hours now takes 15 minutes. I just spec the output; the AI drafts in my style. I do a quick review and ship.
- You direct the AI's behavior — I configure how Athena responds to me. Semantic search gives me contextual, nuanced answers grounded in my own documented principles — not generic advice.
- Zero operational burden — Unlike SaaS products that break at scale, this is a single-user local tool. The complexity is real (1,079+ sessions, vector search, knowledge graphs), but there's no production to break. Real system design, zero ops chaos — ideal for demonstrating engineering depth without the liability.
- Bilateral growth — You evolve with Athena (Ries, 2011). Every insight you learn — about coding, system design, or any domain — gets captured and integrated. The system upgrades alongside you. It's not a static tool; it's a compound knowledge machine that reflects your growth trajectory.
- Quadrant IV: What you don't know you don't know — The highest-value zone (Luft & Ingham, 1955; Rumsfeld, 2002). You can't Google what you don't know exists. Athena surfaces blind spots before they become problems — flagging dependencies you missed, risks you didn't consider, and connections you couldn't see. It's a co-pilot that sees the ditch before the driver does.
"The goal isn't just to answer questions. It's to ask the questions you didn't know to ask."
👉 docs/concepts/Quadrant_IV.md — Deep dive: the 5 mechanisms that make this work
One AI is not enough for life decisions.
This is Athena's biggest unlock: cross-model validation that catches idiosyncratic errors and forces deeper investigation when models disagree.
Important
Both the User AND the AI can be blind. Cross-model consensus ("The Watchmen watching the Watchmen") prevents both parties from getting "high on their own supply."
| Blind Spot Type | Fixed By |
|---|---|
| User Blind Spot | Athena's Internal Routing (The Skeptic) |
| Athena's Blind Spot | External Models (ChatGPT, Grok) |
| Mutual Blindness (Shared Illusion) | Multi-Model Consensus (3/3 Agree = High Confidence) |
Tip
The Watchmen Rule: If 3 rival models (trained on different data, with different biases) all identify the same risk, the probability of it being a hallucination drops to near zero.
flowchart LR
A[You] -->|1. Query| B["Athena<br/>(Claude)"]
B -->|2. Discuss| A
A -->|3. Export Artifact| C["AI #2<br/>Gemini"]
A -->|3. Export Artifact| D["AI #3<br/>ChatGPT"]
A -->|3. Export Artifact| E["AI #4<br/>Grok"]
C -->|4. Red-Team Audit| F[Findings]
D -->|4. Red-Team Audit| F
E -->|4. Red-Team Audit| F
F -->|5. Return| B
B -->|6. Synthesize| G[Final Conclusion]
style A fill:#4a9eff,color:#fff
style B fill:#cc785c,color:#fff
style C fill:#4285f4,color:#fff
style D fill:#10a37f,color:#fff
style E fill:#1da1f2,color:#fff
style G fill:#22c55e,color:#fff
Expose Athena's brain to any MCP-compatible client.
9 tools + 2 resources, dual transport (stdio + SSE), with a full permissioning layer.
# Launch via stdio (for IDE integration)
python -m athena.mcp_server
# Or via SSE (for remote access)
python -m athena.mcp_server --sse --port 8765| Tool | Permission | Description |
|---|---|---|
smart_search |
read | Hybrid RAG search with RRF fusion |
agentic_search |
read | Agentic RAG v2 — multi-step query decomposition with parallel search |
quicksave |
write | Save checkpoint to session log |
health_check |
read | System health audit |
recall_session |
read | Read session log content |
governance_status |
read | Triple-Lock compliance state |
list_memory_paths |
read | Memory directory inventory |
set_secret_mode |
admin | Toggle demo mode (blocks internal tools) |
permission_status |
read | Show access state & tool manifest |
Toggle set_secret_mode(True) for demos or external sharing. Only PUBLIC tools remain accessible; sensitive content is auto-redacted.
👉 Full MCP Documentation — Architecture, permissioning, and IDE configuration
This repo documents Winston's personal Athena instance — 1,079+ sessions, 330 unique protocols, production-tested daily since June 2025.
It's included as a reference, not a prescription. Your instance will reflect your domain, your decisions, your voice.
"Aren't these the same thing?" — No.
OpenClaw (162k ⭐) is a personal AI assistant platform — it gets your AI into 15+ messaging channels (WhatsApp, Telegram, Slack, Discord, iMessage, etc.) with voice, Canvas UI, and mobile apps. It excels at distribution: making the AI accessible everywhere you already communicate.
Athena is a sovereign memory persistency layer — it gives any AI agent a long-term memory that survives across sessions, models, and IDEs. You own the data; the LLM providers supply the intelligence. It excels at depth: compounding knowledge over 1,000+ sessions.
| Dimension | OpenClaw 🦞 | Athena 🏛️ |
|---|---|---|
| Core Focus | Distribution (reach your AI anywhere) | Memory (your AI remembers everything) |
| Channels | 15+ (WhatsApp, Telegram, Slack, etc.) | IDE-native (Antigravity, Cursor, VS Code) |
| Memory Model | Session pruning (context window) | Persistent knowledge graph + vector search |
| Architecture | Gateway → Agent → Channel | .framework/ + .context/ + .agent/ |
| Voice | ✅ Voice Wake + Talk Mode (ElevenLabs) | ❌ Text-only |
| Mobile | ✅ iOS + Android apps | ❌ Desktop IDE only |
| Knowledge Graph | ❌ | ✅ GraphRAG (entities + communities) |
| Protocol Library | Skills (ClawHub registry) | 324 decision frameworks + case studies |
| Session Continuity | Per-channel session history | Cross-session learning (1,000+ sessions deep) |
| Best For | "I want my AI on WhatsApp" | "I want my AI to remember Session 19 in Session 995" |
Tip
They're complementary, not competing. Use OpenClaw as the interface layer (how you reach your AI) and Athena as the memory layer (what your AI knows). You can use both.
MIT License — see LICENSE
Tip
New to AI agents? Start with What Is an AI Agent? → then follow Your First Agent for a 5-minute quickstart.
| Document | What It Shows |
|---|---|
| 🏆 Top 10 Protocols | MCDA-ranked essential protocols for any AI user |
| 🔍 Quadrant IV: Unknown Unknowns | How Athena surfaces blind spots you didn't know existed — 5 compounding mechanisms |
| 📈 Protocol 106: Min-Max Optimization | Maximize Utility while Minimizing CapEx — with Indifference Curve visualization |
| 🔄 User-Driven RSI | The bilateral loop: how you and AI improve together |
| 📋 Spec Sheet | Project specification: role, scope, constraints, and acceptance criteria |
| 👤 About Me | Career narrative and professional depth |
| 📊 Benchmarks | Real performance metrics (boot time, search latency, token economics) |
| 📚 References | APA-formatted academic citations for all frameworks used |
| 🎬 Demo Guide | Live walkthrough of the system in action |
| ⚡ Adaptive Latency (RETO) | Macro-Robust, Micro-Efficient — how the OS stays safe while the AI sprints |
📁 Case Studies
| Case Study | What It Demonstrates |
|---|---|
| Boot Optimization | 85% boot time reduction via caching & parallelization |
| Search Quality | RRF fusion for hybrid semantic search |
| Protocol Enforcement | Governance engine for compliance |
| Vibe Coding | Zero-point UI development on a budget |
| Silent Partner | BCM analysis for corporate strategy |
| Auto-Blog | Multi-agent n8n workflow for content velocity |
🔒 Security Model
| Mode | Where Data Lives | Best For |
|---|---|---|
| Cloud | Supabase (your project) | Cross-device access, collaboration |
| Local | Your machine only | Sensitive data, air-gapped environments |
| Hybrid | Local files + cloud embeddings | Best of both (embeddings only leave machine) |
Sensitive data? Keep it local. The
athenaSDK supports local vector stores (ChromaDB, LanceDB) for users who don't want data leaving their machine. See docs/LOCAL_MODE.md.
| Component | Sends Raw Text? | Sends Embeddings? | Destination |
|---|---|---|---|
| Embedding API | Yes (text chunks) | — | Google Cloud |
| LLM API | Yes (prompts) | — | Anthropic (Claude) |
| Supabase | No | Yes (vectors only) | Your Supabase project |
- Supabase Keys: Use
SUPABASE_ANON_KEYfor client-side operations. Never exposeSUPABASE_SERVICE_ROLE_KEYin code or logs. - Row-Level Security: Enable RLS on Supabase tables. See SECURITY.md for policy templates.
- Agentic Safety: If using an agentic IDE with filesystem access, restrict the agent's working directory. Never grant access to
~/.ssh,.envfiles, or git credentials.
Supabase is not just a search layer — it's a backup of all indexed memories. If local files are lost, the vector database enables full recovery.
| Failure Scenario | Recovery Path |
|---|---|
| Local disk failure | Pull from Supabase embeddings → reconstruct Markdown |
| Accidental deletion | Re-index from cloud → restore local files |
| Session corruption | Replay from session_logs table |
Philosophy: Cloud is not "home" — it's insurance.
Why Redundancy Matters: In system design, redundancy is the intentional duplication of critical components to increase reliability. Athena follows this principle:
- Primary: Local Markdown files (git-versioned, human-readable)
- Secondary: Supabase vector embeddings (cloud-native, searchable)
- Tertiary: Session logs with timestamps (audit trail)
This isn't over-engineering — it's survival. Platform APIs change. Local disks fail. The only hedge is strategic duplication.
⚙️ Prerequisites (API Keys)
- Python 3.10+
- Supabase project with pgvector enabled (setup guide) — or use local mode
- API keys in
.env:
# Required
SUPABASE_URL=https://your-project.supabase.co
SUPABASE_ANON_KEY=your-anon-key # NOT service_role key
ANTHROPIC_API_KEY=your-anthropic-key # For Claude reasoning
# Optional (for trilateral feedback with multiple LLMs)
GOOGLE_API_KEY=your-google-api-key
OPENAI_API_KEY=your-openai-keycp .env.example .env
# Add your keys to .env🛠️ Tech Stack & Architecture
| Layer | Technology | Purpose |
|---|---|---|
| SDK | athena Python package |
Core search, reranking, memory |
| Reasoning | Claude Opus 4.6 (primary) | Main reasoning engine |
| IDE | Antigravity (supports Claude, Gemini, GPT) | Agentic development environment |
| Local Models | Ollama, LM Studio (any model) | Model-agnostic — memory is in files, not the model |
| Embeddings | text-embedding-004 (768-dim) |
Google embedding model |
| GraphRAG | NetworkX + Leiden + ChromaDB | Knowledge graph |
| Memory | Supabase + pgvector or local (ChromaDB) | Vector database |
| Knowledge Store | Markdown files (git-versioned) | Human-readable, locally owned |
| Reasoning Engine | Graph of Thoughts (GoT) | DAG-based planning with aggregation/synthesis |
| Query Routing | CognitiveRouter | Adaptive latency based on query complexity |
| Budget Enforcement | BudgetGatekeeper | Real-time token/tool/cost limit enforcement |
┌─────────────────────────────────────────────────────────────────────────┐
│ │
│ (1) /start ──► Retrieve Context ──► (2) Work ──► (3) /end │
│ ▲ │ │
│ │ ▼ │
│ └───── (5) Next Session ◄── Embed ◄── (4) Extract & Store │
│ │
└─────────────────────────────────────────────────────────────────────────┘
Think of it like Git, but for conversations. Each session builds on the last. Important decisions get captured, indexed, and recoverable.
| Feature | How It Works |
|---|---|
/start boot |
Loads identity + retrieves relevant context from long-term memory |
/end commit |
Summarizes session, extracts decisions, saves to knowledge store |
| Hybrid search | Fuses Canonical + GraphRAG + Tags + Vectors + Filenames via RRF |
| MCP Server | 9 tools via Model Context Protocol — any MCP client can connect |
| Permissioning | 4 capability levels + 3 sensitivity tiers + Secret Mode (demo/external) |
| Evaluator Gate | 50-query regression suite: MRR@5 = 0.44, Hit@5 = 52% |
| Cross-encoder reranking | Refines top results with sentence-transformers |
| Protocol library | 330 unique protocols (75 curated in starter pack) |
Athena-Public/
├── src/athena/ # SDK package (pip installable)
│ ├── core/ # Config, models, governance, permissions
│ ├── tools/ # Search, reranker, latency
│ ├── memory/ # Vector DB interface
│ └── mcp_server.py # MCP Tool Server (9 tools, 2 resources)
├── examples/
│ ├── quickstart/ # Runnable demos
│ ├── scripts/ # Automation scripts
│ ├── protocols/ # Thinking patterns (starter pack included)
│ ├── workflows/ # Slash commands
│ └── templates/ # Starter templates
├── docs/ # Deep documentation
├── community/ # Contributing, roadmap
├── pyproject.toml # Modern packaging
└── .env.example # Environment template
📖 Key Concepts & Workflows
- docs/ARCHITECTURE.md — How the system is designed
- docs/GRAPHRAG.md — Knowledge graph layer (community detection + entity search)
- docs/VECTORRAG.md — Semantic memory implementation
- docs/GETTING_STARTED.md — Build your own
- examples/concepts/adaptive_latency.md —
/start,/think,/ultrathink - docs/GLOSSARY.md — Key terms and definitions
| Command | Description |
|---|---|
/start |
Boot system, load identity |
/end |
Close session, commit to memory |
/think |
Deep reasoning mode |
/ultrathink |
Maximum depth analysis |
/brief |
Pre-prompt clarification protocol |
/refactor |
Workspace optimization |
/research |
Multi-source web research |
/plan |
Structured planning with pre-mortem |
/save |
Mid-session checkpoint |
/vibe |
Ship at 70%, iterate fast |
👉 Full Workflow Documentation — Detailed guide to all 12 workflows
See .agent/workflows/ for source files.
📋 Changelog
- v8.5.0 (Feb 12 2026): Phase 1 Complete — MCP Tool Server (9 tools, 2 resources), Permissioning Layer (4 levels + secret mode), Search MRR +105% (0.21→0.44), Evaluator Gate (50 queries). SDK v2.0.0.
- v8.3.1 (Feb 11 2026): Viral Validation Release — 654K+ Reddit views, 1,660+ upvotes, 5,300+ shares. #1 All-Time r/ChatGPT, #2 All-Time r/GeminiAI. Three-Phase Token Budget: robustness at
/start+/end, Adaptive Latency in-between. - v8.2.1 (Feb 9 2026): Metrics Sync — Fixed
batch_audit.pyautomation, linked orphan files, reconciled tech debt, 8,079 tags indexed
👉 Full Changelog — Complete version history from v1.0.0 (Dec 2025)
For the full documentation, case studies, and deep dives, see docs/.
