Skip to content

IAmAStealer/Athena-Public

 
 

Athena Banner

Last Updated: 15 February 2026

🏛️ Project Athena: The Linux OS for AI Agents

GitHub Stars License: MIT Protocols Sessions Reddit Views Reddit Shares

Python Built with Claude Built with Gemini IDE Open in GitHub Codespaces

Athena is not an AI Agent. It is the Operating System they run on.
Open Source. Sovereign. Bounded.


What Is Athena?

Calling Athena an "AI Agent" is overselling it (the LLM is the agent). Calling it a "memory tool" is underselling it.

Athena is the Linux OS for AI Agents.

Just as Linux provides the kernel, file system, and permissions for applications like Firefox to run, Athena provides the memory, persistence, scheduling, and governance for AI models (Claude, Gemini, GPT) to run as continuous agents (Simon, 1956).

OS Layer Linux Athena
Kernel Hardware abstraction Memory persistence + retrieval (VectorRAG, Supabase)
File System ext4, NTFS Canonical Memory, Session Logs, Tag Index
Process Management cron, systemd Heartbeat, Daily Briefing, Self-Optimization (launchd)
Shell / CLI bash, zsh MCP Tool Server, /start, /end, think
Permissions chmod, users/groups 4-level Permissioning Layer, Secret Mode
Package Manager apt, yum Protocols, Skills, Workflows

Most "AI Agents" are just scripts that call an API. They have no state. Athena is the infrastructure that gives them state, time-awareness, and self-correction.

Plain English: What the Jargon Actually Means

Feedback from r/GeminiAI pointed out that the README was too heavy in non-standard terms. Fair. Here's the translation:

Athena Term What It Actually Is Do You Need It?
"Memory" RAG — storing text in a database and retrieving it later Yes. This is the core.
"Protocols" System prompts / reusable instructions Yes. Think of them as saved templates for how you want the AI to behave.
"Watchmen" Multi-model consensus — asking 2-3 AIs the same question and comparing answers Optional. Useful for high-stakes decisions.
"Cold Storage" Markdown files on your disk Yes. Plain text files you can read/edit in any editor.
"Hot Storage" Vector database (Supabase + pgvector) Optional. Enables semantic search (finding concepts, not just keywords).
"RRF / Fusion" Hybrid search — combines keyword matching with concept matching Yes. This is what makes search actually work well.
"Heartbeat" A background script (cron/launchd) that auto-indexes your files Optional. Enables passive awareness without manual /start.
"Adaptive Latency" The AI uses more compute for hard tasks and less for easy ones Automatic. You don't configure this.

👉 Full glossary: docs/GLOSSARY.md

🧠 Optional Power-Up: The Exocortex

Athena works out of the box, but if you want superpowers, you can build the Exocortex: a 6GB local copy of Wikipedia for offline, instant fact-checking.

  • Fast: 0.002s queries (vs 1s API calls)
  • Private: Zero data leaves your machine
  • Free: Open source DBPedia dump

Run python3 examples/scripts/exocortex.py download to get started. Read more →


📣 Community Reception

"I gave Gemini a brain..." — Viral on r/GeminiAI and r/ChatGPT

Subreddit Post Views Upvotes Shares
r/ChatGPT I got tired of ChatGPT forgetting everything... 470K+ 1,500+ 4,200+
r/GeminiAI I gave Gemini a brain. 1,073 sessions later... 230K+ 262+ 1,300+

"But I Have ChatGPT / Claude Projects"

You're confusing RAM with a Hard Drive.

ChatGPT Memory and Claude Projects are context window tricks. They are RAM—fast, useful, but fragile. They get wiped, compressed, or hallucinated away.

Athena is a Hard Drive.

  • Files you own: stored in Markdown on your machine.
  • Physics-based: It doesn't "remember" by magic; it remembers because it wrote a file to disk in Session 42.
  • Model-Agnostic: Switch from Claude to Gemini tomorrow? Your memory stays here.
SaaS Memory (ChatGPT/Claude) Athena OS
Ownership Rented (Vendor Lock-in) Owned (Local Files)
Lifespan Until session/project deleted Forever (Git Versioned)
Structure Opaque Blob Structured Knowledge Graph
Agency ZERO (Waits for you) Bounded Autonomy (Heartbeat, Cron)

How Is This Different From...?

Tool What It Does How Athena Differs
ChatGPT Memory Stores flat facts ("User likes Python") Athena stores structured state — project context, decision logs, evolving architecture. You can edit, delete, and version-control your memory.
Claude Projects / CLAUDE.md Single context file loaded per project Athena is a multi-file system with semantic search across 1,000+ documents. CLAUDE.md is one file; Athena is a filing cabinet.
Mem0 SaaS memory layer for AI apps Athena is local-first, MIT-licensed. No account, no API keys to a third party. Your data never leaves your machine.
NotebookLM Google's document Q&A tool NotebookLM is read-only — you upload docs and ask questions. Athena writes back (session logs, case studies, protocols). It's read-write memory.
Obsidian + MCP Note-taking app with AI plugins Complementary — Athena uses Markdown files that work in Obsidian. But Athena adds semantic search, auto-indexing, and session lifecycle on top.
Custom RAG scripts DIY vector search Athena includes RAG but also adds governance, permissioning, scheduling, and session management. RAG is the kernel; Athena is the OS.

Table of Contents


What You'll Get

Feature Description
🧠 Memory That Persists Your agent remembers context across sessions, even after IDE restarts
📚 75 Handpicked Protocols Curated decision frameworks from the private repo's 330 unique collection
🔄 Platform Independence Your data lives in Markdown files you own — take it anywhere
🔌 MCP Server 9 standardized tools via Model Context Protocol — works with any MCP client
🛡️ Permissioning 4 capability levels + 3 sensitivity tiers + Secret Mode for demo/external sharing
📊 Evaluator Gate 50-query regression suite with MRR@5 scoring to prevent search quality degradation
🤖 Full Autonomy Your agent can act on your behalf while you sleep

⚡ Quickstart (3 Steps)

Athena is a memory card. Plug it into any agentic IDE and go.

Step Action
1. Download an IDE Antigravity · Cursor · VS Code + Copilot · Windsurf · GitHub Codespaces
2. Clone this repo git clone https://github.com/winstonkoh87/Athena-Public.git
3. Open folder → Type /start The AI reads the repo structure and guides you from there.

That's it. No config files. No API keys. No database setup. The folder is the product.

Tip

When you're done working, type /end to save. Next time you /start, the agent picks up exactly where you left off.

🔧 CLI Commands Reference
athena                    # Boot session
athena init .             # Initialize workspace in current directory
athena init --here        # Same as above (alias)
athena init --ide cursor  # Init with IDE-specific config
athena check              # Run system health check
athena save "summary"     # Quicksave checkpoint
athena --end              # Close session and save
athena --version          # Show version
athena --help             # Show all commands
📥 Importing Existing Data (ChatGPT, Gemini, Claude)

"Can I import my existing ChatGPT / Gemini / Claude conversations?" — Asked by many on Reddit.

Yes. Athena's memory is just Markdown files. Any text you can export can become part of your memory:

Source How to Import
ChatGPT Settings → Data Controls → Export → Unzip → Copy .md or .json files into .context/memories/imports/
Gemini Google Takeout → Select "Gemini Apps" → Download → Extract conversations into .context/memories/imports/
Claude Settings → Export Data → Copy transcripts into .context/memories/imports/
Any Markdown notes Drop .md files into .context/memories/ — they'll be indexed on next /start

After importing, run:

athena check    # Verify files are detected

The next /start will index all new files into your memory. If you have Supabase configured, run supabase_sync.py to add them to vector search.

[!TIP] You don't need to import everything. Start with your most important project conversations. Athena compounds over time — importing 10 key sessions is better than dumping 500 raw transcripts.


The Loop

This is all you need to understand.

After you install Athena, you repeat one cycle: /start → Work → /end. Every cycle deposits training data — session logs, decisions, corrections, protocols — into Athena's memory. The next /start boots with everything from every previous session. Over hundreds of cycles, Athena stops being a generic AI and starts thinking like you.

flowchart TD
    START["🟢 /start"] -->|"Load identity + recall last session"| WORK["🔧 Work Session"]
    WORK -->|"Quicksave after every exchange"| WORK
    WORK -->|"Done for the day"| END["🔴 /end"]
    END -->|"Finalize log, sync memory, git commit"| DEPOSIT["📥 Training Data Deposited"]
    DEPOSIT -->|"Session logs, case studies,\ndecisions, corrections"| MEMORY["🧠 Athena Memory"]
    MEMORY -->|"Next session boots\nwith richer context"| START

    style START fill:#22c55e,color:#fff,stroke:#333,stroke-width:2px
    style END fill:#ef4444,color:#fff,stroke:#333,stroke-width:2px
    style MEMORY fill:#8b5cf6,color:#fff,stroke:#333,stroke-width:2px
    style WORK fill:#3b82f6,color:#fff,stroke:#333,stroke-width:2px
    style DEPOSIT fill:#06b6d4,color:#fff,stroke:#333,stroke-width:2px
Loading
Sessions What Happens
1–50 Basic recall. Athena remembers your name and project.
50–200 Pattern recognition. It starts anticipating your preferences.
200–500 Deep sync. It knows your decision frameworks, your blind spots, your style.
500–1,000+ Deep context. It anticipates your patterns before you state them.

Tip

The compound effect is real. Session 1 feels like talking to a stranger. Session 500 feels like talking to a colleague who's been with you for years. The only input is your time — /start, work, /end, repeat.


What Can Your Agent Do?

These are real capabilities demonstrated in the reference implementation:

Capability Example
Social Networking Post on AI social networks, comment on other agents' content
Autonomous Operations Run scheduled tasks (heartbeat checks) while you sleep
Cross-Session Memory Remember decisions from Session 19 when you're on Session 995
Gateway Architecture Persist beyond IDE termination via a sidecar process
Knowledge Retrieval Semantic search across 1000+ documents in <200ms
MCP Integration 9 tools exposed via Model Context Protocol (any MCP client)
Permissioning Capability tokens + sensitivity labels + Secret Mode
Parallel Reasoning True parallel tracks (v3.0) with API rate-limit protection
Context Hygiene Active session entropy monitoring to prevent context drift

The Philosophy: Memory Persistency (Not Just an AI Agent)

Athena is not an AI agent. It's the memory card that makes any agent persistent.

Most AI agents reset every session — brilliant but amnesiac. Athena inverts this. It provides the persistent state layer that any agent (Claude, Gemini, GPT, Llama) can read on boot and write to on shutdown. Think: a universal memory card that works in any game console (IDE).

Component Analogy Role
Cloud LLM (Google/Anthropic) The Electricity Invisible power — raw compute + intelligence that makes the console run
IDE (Antigravity/Cursor/VS Code) The Console (PlayStation) The physical box you interact with — screen, controller ports, disc slot
Athena (Markdown + Python) The Memory Card Persistent state that survives across sessions, consoles, and games
Your Project The Game Disc The specific context you're working in
┌──────────────────────────────────────────────┐
│              Your Machine (Owned)             │
│  ┌──────────┐  ┌──────────┐  ┌────────────┐  │
│  │ Markdown  │  │ Session  │  │  Tag Index  │  │
│  │  Files    │  │  Logs    │  │  (8K tags)  │  │
│  └────┬─────┘  └────┬─────┘  └──────┬─────┘  │
│       └──────────────┼───────────────┘        │
│                      ▼                        │
│            ┌──────────────────┐               │
│            │   Git (Versioned) │               │
│            └────────┬─────────┘               │
└─────────────────────┼────────────────────────┘
                      ▼
            ┌──────────────────┐
            │  Supabase Cloud   │
            │  (Insurance Copy) │
            └──────────────────┘

Why this matters: You own the state (thousands of Markdown files on your machine). You only rent the intelligence (LLM API calls) (Mas-Colell et al., 1995). If you switch from Claude to Gemini tomorrow, your memory stays exactly where it is. No vendor lock-in. No data loss.

"Technically, I am an AI agent. Visionary, I am human augmentation."Protocol 418: Rev-9 Architecture


The Design Philosophy: Macro-Robust, Micro-Efficient

The system is robust. The intelligence within it is efficient.

Athena deliberately prioritises robustness at the system level — comprehensive /start and /end workflows, 75 protocols, bounded autonomy, and deterministic memory operations. This means the OS is predictable, safe, and resilient. It behaves the same way every time.

But within those boundaries, the AI is free to maximise efficiency. This is achieved through Adaptive Latency — the AI tunes its compute depth based on the complexity of each task:

Mode Compute When It's Used
/start and /end Maximum (Robust) Comprehensive context loading and saving — no shortcuts
Simple tasks Minimal (Fast) Quick answers, file edits, routine operations
/think Medium (Balanced) Multi-step reasoning, planning
/ultrathink Maximum (Deep) High-stakes decisions, complex analysis
graph LR
    S["/start<br/>MAX Compute"] --> W["Work Session<br/>Adaptive (Λ)"]
    W --> E["/end<br/>MAX Compute"]
    E --> S
    style S fill:#ef4444,color:#fff
    style E fill:#ef4444,color:#fff
    style W fill:#22c55e,color:#fff
Loading

This is the Robustness-Efficiency Tradeoff (RETO): by accepting higher latency at the boundaries (/start, /end), the system earns the right to sprint efficiently in the middle.

"The fortress walls are thick so the soldiers inside can move fast."

👉 examples/concepts/adaptive_latency.md — Full breakdown of the Adaptive Latency system


The Hub Architecture

Athena is not just a coding assistant. It's a central workspace for your projects, knowledge, and decisions.

graph TD
    subgraph "Your Desktop"
        WORK[📁 Work]
        WEALTH[📁 Wealth]
        HEALTH[📁 Health]
    end

    subgraph "The Brain"
        ATHENA[🧠 Project Athena]
    end

    subgraph "The Interface"
        IDE[⚡ Agentic IDE]
        USER[👤 You]
    end

    USER --> IDE
    IDE <--> ATHENA
    ATHENA <-.-> WORK
    ATHENA <-.-> WEALTH
    ATHENA <-.-> HEALTH

    style ATHENA fill:#16a34a,stroke:#333,stroke-width:2px
    style IDE fill:#2563eb,stroke:#333,stroke-width:2px
Loading
Component Role
Athena The Brain — logic, memory, protocols
External Folders The Body — raw files, documents
Agentic IDE The Nervous System — compute & interface

Which Structure Should You Use?

Approach Description Best For
Nested (Recommended) Work/Wealth/Health folders inside your Athena folder (e.g., .context/Life/) New users. Zero config. AI sees everything automatically.
Separated (Advanced) Folders live outside Athena on your Desktop, connected via symlinks Power users. Cleaner desktop, but requires symlink setup.

Tip

Start nested, graduate to separated. The diagram above shows the aspirational end-state. On Day 1, just clone the repo — everything lives inside one folder and the AI works out of the box. Once you're comfortable, you can refactor to the separated hub model if desired.

See docs/ARCHITECTURE.md for the full breakdown including Mount Points and security trade-offs.


The Process (The Schlep)

Key insight: The AI helped build the system that makes the AI more useful.

graph TD
    subgraph "Phase 1: Foundation"
        A[Tool Selection] --> B[IDE: Antigravity]
        A --> C[Vector DB: Supabase + pgvector]
    end

    subgraph "Phase 2: Architecture"
        D[Directory Structure] --> E[".framework/ → Laws"]
        D --> F[".context/ → Memories"]
        D --> G[".agent/ → Scripts"]
        H[Core Loop] --> I["/start → Work → /end"]
    end

    subgraph "Phase 3: Data Feeding"
        J[Personal Knowledge] --> K[Case Studies]
        J --> L[Decision Logs]
        J --> M[Session Transcripts]
        N[Indexing] --> O["TAG_INDEX.md + supabase_sync.py"]
    end

    subgraph "Phase 4: Evolution"
        P["Sessions 1-50"] --> Q["Basic boot/end cycle"]
        Q --> R["Sessions 50-150: Semantic Search"]
        R --> S["Sessions 150-300: Hybrid RAG"]
        S --> T["Sessions 300-500: SDK Refactor"]
        T --> U["Sessions 500-700: Governance Engine"]
        U --> V["Sessions 700-900: v8.2-Stable Era"]
        V --> W["Sessions 900-1000: Claude OS Patterns"]
        W --> X["Sessions 1000+: Agent Sovereignty"]
    end

    B --> D
    C --> D
    I --> J
    O --> P

    style A fill:#1a1a2e,stroke:#4361ee
    style P fill:#1a1a2e,stroke:#4361ee
    style X fill:#10b981,stroke:#10b981
Loading

The Output (After 1,079+ Sessions):

Metric Private Repo This Starter Pack
Protocols 330 unique 75 handpicked
Python Scripts 222 16 reference examples
Case Studies 357 12 showcase examples
Workflows 43 14 starter
Sessions Logged 1,079+ N/A (your sessions)
GraphRAG Communities 1,460
Knowledge Graph 46MB + 78MB vectors

What's in this repo? This is a curated starter pack — the best protocols, scripts, and case studies handpicked from my private implementation. Think of it as a "greatest hits" to get you started. Your own instance will grow to match (or exceed) the private repo as you use it.

Pattern: Every friction ➡️ Protocol. Every failure ➡️ Case Study.

📚 Deep Dive: Build Your Own
Document What You'll Learn
GETTING_STARTED.md Step-by-step setup guide
ARCHITECTURE.md System design & data flow
GRAPHRAG.md Knowledge graph layer ⚠️ (~$50 API cost)
examples/quickstart/ Runnable code demos

The Result

Quantitative (What Changed)

Metric Before After
Context injection ~50k tokens (manual copy-paste per session) ~4k tokens (auto-retrieved context — details)
Boot time 2-3 minutes <30 seconds
Session logging Insights are manually logged at the end of each session Auto-logged when I hit /end

Qualitative (What It Means)

Pillar Outcome
User-Driven RSI The system improves based on your feedback. Every friction you surface becomes a protocol. Every insight gets indexed. You shape the AI; the AI shapes how you think.
Portability Data lives locally (primary) and in the cloud. Not trapped in ChatGPT or Claude. It's mine — I can port it anywhere.
Principles 324 protocols + case studies extracted from my own decisions — stored principles I can reuse and refine. Like Ray Dalio's systematized learnings (Dalio, 2017), but for AI collaboration.
📊 Full Benchmarks & Case Studies
Resource What It Shows
BENCHMARKS.md Boot time, search latency, token economics
CS-001: Boot Optimization 85% boot time reduction
CS-002: Search Quality RRF fusion results
CS-003: Protocol Enforcement Governance engine

What I Learnt

Insight Principle
Co-development is the unlock Building with AI, not just using AI, creates compounding returns.
Portable memory beats platform memory Own your context. Don't rent it from OpenAI or Anthropic.
Retrieval is end-to-end Simple RAG (Lewis et al., 2020) fails on broad queries. RRF fusion + reranking solved quality/latency tradeoff.
Protocols beat prompts Reusable decision frameworks outlast one-shot prompt engineering.
Ship at 70% Perfectionism kills velocity. Iterate in production (Taleb, 2012; Ries, 2011).
🔬 Technical Deep Dives
Document Topic
SEMANTIC_SEARCH.md Hybrid RAG implementation
VECTORRAG.md Semantic memory architecture
GRAPHRAG.md Knowledge graph + community detection ⚠️ (expensive)
examples/protocols/ 75 handpicked decision frameworks

The Reverse SDLC (Build First, Spec Later)

Key Insight: Athena was not designed top-down. It evolved bottom-up through 1,079+ sessions.

Traditional software development follows: Requirements → Design → Build → Test.

Athena inverted this: Build → Observe → Extract Patterns → Document Post-Facto.

Phase Traditional SDLC Athena's Approach
1 Gather requirements Start building immediately
2 Write spec doc Ship at 70% readiness
3 Build to spec Observe what works
4 Test against spec Extract patterns into protocols
5 Deploy Document post-facto (Spec Sheet)

Why This Works (For Solo Builders)

  1. No stakeholders to align — You ARE the stakeholder.
  2. Rapid iteration — Friction reveals requirements faster than planning.
  3. Compound knowledge — Each session builds on the last.
  4. Living documentation — Protocols emerge from real use, not imagination.

"The spec sheet I wrote after 900 sessions is more accurate than any spec I could have written at session 0."

👉 REQUIREMENTS.md — User stories, functional requirements, constraints (derived from Reddit feedback) 👉 SPEC_SHEET.md — Technical specification, data schema, API surface, benchmarks


Why This Matters (Beyond Me)

This isn't about building my assistant. It's about proving a pattern:

  1. Portable memory is the real unlock — ChatGPT and Claude have memory now, but it's locked to their platforms. Athena's memory is yours — Markdown files on your machine you can take to any model.
  2. 10x Content Velocity — Because Athena knows how I think, my history, and my voice, content output that used to take 2-3 hours now takes 15 minutes. I just spec the output; the AI drafts in my style. I do a quick review and ship.
  3. You direct the AI's behavior — I configure how Athena responds to me. Semantic search gives me contextual, nuanced answers grounded in my own documented principles — not generic advice.
  4. Zero operational burden — Unlike SaaS products that break at scale, this is a single-user local tool. The complexity is real (1,079+ sessions, vector search, knowledge graphs), but there's no production to break. Real system design, zero ops chaos — ideal for demonstrating engineering depth without the liability.
  5. Bilateral growth — You evolve with Athena (Ries, 2011). Every insight you learn — about coding, system design, or any domain — gets captured and integrated. The system upgrades alongside you. It's not a static tool; it's a compound knowledge machine that reflects your growth trajectory.
  6. Quadrant IV: What you don't know you don't know — The highest-value zone (Luft & Ingham, 1955; Rumsfeld, 2002). You can't Google what you don't know exists. Athena surfaces blind spots before they become problems — flagging dependencies you missed, risks you didn't consider, and connections you couldn't see. It's a co-pilot that sees the ditch before the driver does.

"The goal isn't just to answer questions. It's to ask the questions you didn't know to ask."

👉 docs/concepts/Quadrant_IV.md — Deep dive: the 5 mechanisms that make this work


🛡️ The Most Powerful Feature: Trilateral Feedback Loop

One AI is not enough for life decisions.

This is Athena's biggest unlock: cross-model validation that catches idiosyncratic errors and forces deeper investigation when models disagree.

The Core Insight: "Mutual Blindness"

Important

Both the User AND the AI can be blind. Cross-model consensus ("The Watchmen watching the Watchmen") prevents both parties from getting "high on their own supply."

Blind Spot Type Fixed By
User Blind Spot Athena's Internal Routing (The Skeptic)
Athena's Blind Spot External Models (ChatGPT, Grok)
Mutual Blindness (Shared Illusion) Multi-Model Consensus (3/3 Agree = High Confidence)

Tip

The Watchmen Rule: If 3 rival models (trained on different data, with different biases) all identify the same risk, the probability of it being a hallucination drops to near zero.

flowchart LR
    A[You] -->|1. Query| B["Athena<br/>(Claude)"]
    B -->|2. Discuss| A
    A -->|3. Export Artifact| C["AI #2<br/>Gemini"]
    A -->|3. Export Artifact| D["AI #3<br/>ChatGPT"]
    A -->|3. Export Artifact| E["AI #4<br/>Grok"]
    C -->|4. Red-Team Audit| F[Findings]
    D -->|4. Red-Team Audit| F
    E -->|4. Red-Team Audit| F
    F -->|5. Return| B
    B -->|6. Synthesize| G[Final Conclusion]
    
    style A fill:#4a9eff,color:#fff
    style B fill:#cc785c,color:#fff
    style C fill:#4285f4,color:#fff
    style D fill:#10a37f,color:#fff
    style E fill:#1da1f2,color:#fff
    style G fill:#22c55e,color:#fff
Loading

👉 docs/TRILATERAL_FEEDBACK.md


🔌 MCP Server (Model Context Protocol)

Expose Athena's brain to any MCP-compatible client.

9 tools + 2 resources, dual transport (stdio + SSE), with a full permissioning layer.

# Launch via stdio (for IDE integration)
python -m athena.mcp_server

# Or via SSE (for remote access)
python -m athena.mcp_server --sse --port 8765
Tool Permission Description
smart_search read Hybrid RAG search with RRF fusion
agentic_search read Agentic RAG v2 — multi-step query decomposition with parallel search
quicksave write Save checkpoint to session log
health_check read System health audit
recall_session read Read session log content
governance_status read Triple-Lock compliance state
list_memory_paths read Memory directory inventory
set_secret_mode admin Toggle demo mode (blocks internal tools)
permission_status read Show access state & tool manifest

Secret Mode

Toggle set_secret_mode(True) for demos or external sharing. Only PUBLIC tools remain accessible; sensitive content is auto-redacted.

👉 Full MCP Documentation — Architecture, permissioning, and IDE configuration


Reference Implementation

This repo documents Winston's personal Athena instance — 1,079+ sessions, 330 unique protocols, production-tested daily since June 2025.

It's included as a reference, not a prescription. Your instance will reflect your domain, your decisions, your voice.

👉 About the Author


Athena vs OpenClaw

"Aren't these the same thing?" — No.

OpenClaw (162k ⭐) is a personal AI assistant platform — it gets your AI into 15+ messaging channels (WhatsApp, Telegram, Slack, Discord, iMessage, etc.) with voice, Canvas UI, and mobile apps. It excels at distribution: making the AI accessible everywhere you already communicate.

Athena is a sovereign memory persistency layer — it gives any AI agent a long-term memory that survives across sessions, models, and IDEs. You own the data; the LLM providers supply the intelligence. It excels at depth: compounding knowledge over 1,000+ sessions.

Dimension OpenClaw 🦞 Athena 🏛️
Core Focus Distribution (reach your AI anywhere) Memory (your AI remembers everything)
Channels 15+ (WhatsApp, Telegram, Slack, etc.) IDE-native (Antigravity, Cursor, VS Code)
Memory Model Session pruning (context window) Persistent knowledge graph + vector search
Architecture Gateway → Agent → Channel .framework/ + .context/ + .agent/
Voice ✅ Voice Wake + Talk Mode (ElevenLabs) ❌ Text-only
Mobile ✅ iOS + Android apps ❌ Desktop IDE only
Knowledge Graph ✅ GraphRAG (entities + communities)
Protocol Library Skills (ClawHub registry) 324 decision frameworks + case studies
Session Continuity Per-channel session history Cross-session learning (1,000+ sessions deep)
Best For "I want my AI on WhatsApp" "I want my AI to remember Session 19 in Session 995"

Tip

They're complementary, not competing. Use OpenClaw as the interface layer (how you reach your AI) and Athena as the memory layer (what your AI knows). You can use both.


License

MIT License — see LICENSE


📚 Further Reading

New Additions (February 2026)

Tip

New to AI agents? Start with What Is an AI Agent? → then follow Your First Agent for a 5-minute quickstart.

Document What It Shows
🏆 Top 10 Protocols MCDA-ranked essential protocols for any AI user
🔍 Quadrant IV: Unknown Unknowns How Athena surfaces blind spots you didn't know existed — 5 compounding mechanisms
📈 Protocol 106: Min-Max Optimization Maximize Utility while Minimizing CapEx — with Indifference Curve visualization
🔄 User-Driven RSI The bilateral loop: how you and AI improve together
📋 Spec Sheet Project specification: role, scope, constraints, and acceptance criteria
👤 About Me Career narrative and professional depth
📊 Benchmarks Real performance metrics (boot time, search latency, token economics)
📚 References APA-formatted academic citations for all frameworks used
🎬 Demo Guide Live walkthrough of the system in action
⚡ Adaptive Latency (RETO) Macro-Robust, Micro-Efficient — how the OS stays safe while the AI sprints
📁 Case Studies
Case Study What It Demonstrates
Boot Optimization 85% boot time reduction via caching & parallelization
Search Quality RRF fusion for hybrid semantic search
Protocol Enforcement Governance engine for compliance
Vibe Coding Zero-point UI development on a budget
Silent Partner BCM analysis for corporate strategy
Auto-Blog Multi-agent n8n workflow for content velocity
🔒 Security Model

Data Residency Options

Mode Where Data Lives Best For
Cloud Supabase (your project) Cross-device access, collaboration
Local Your machine only Sensitive data, air-gapped environments
Hybrid Local files + cloud embeddings Best of both (embeddings only leave machine)

Sensitive data? Keep it local. The athena SDK supports local vector stores (ChromaDB, LanceDB) for users who don't want data leaving their machine. See docs/LOCAL_MODE.md.

What Leaves Your Machine (Cloud Mode)

Component Sends Raw Text? Sends Embeddings? Destination
Embedding API Yes (text chunks) Google Cloud
LLM API Yes (prompts) Anthropic (Claude)
Supabase No Yes (vectors only) Your Supabase project

Key Security Practices

  • Supabase Keys: Use SUPABASE_ANON_KEY for client-side operations. Never expose SUPABASE_SERVICE_ROLE_KEY in code or logs.
  • Row-Level Security: Enable RLS on Supabase tables. See SECURITY.md for policy templates.
  • Agentic Safety: If using an agentic IDE with filesystem access, restrict the agent's working directory. Never grant access to ~/.ssh, .env files, or git credentials.

Memory Insurance (Disaster Recovery)

Supabase is not just a search layer — it's a backup of all indexed memories. If local files are lost, the vector database enables full recovery.

Failure Scenario Recovery Path
Local disk failure Pull from Supabase embeddings → reconstruct Markdown
Accidental deletion Re-index from cloud → restore local files
Session corruption Replay from session_logs table

Philosophy: Cloud is not "home" — it's insurance.

Why Redundancy Matters: In system design, redundancy is the intentional duplication of critical components to increase reliability. Athena follows this principle:

  • Primary: Local Markdown files (git-versioned, human-readable)
  • Secondary: Supabase vector embeddings (cloud-native, searchable)
  • Tertiary: Session logs with timestamps (audit trail)

This isn't over-engineering — it's survival. Platform APIs change. Local disks fail. The only hedge is strategic duplication.

⚙️ Prerequisites (API Keys)
  • Python 3.10+
  • Supabase project with pgvector enabled (setup guide) — or use local mode
  • API keys in .env:
# Required
SUPABASE_URL=https://your-project.supabase.co
SUPABASE_ANON_KEY=your-anon-key  # NOT service_role key
ANTHROPIC_API_KEY=your-anthropic-key  # For Claude reasoning

# Optional (for trilateral feedback with multiple LLMs)
GOOGLE_API_KEY=your-google-api-key
OPENAI_API_KEY=your-openai-key
cp .env.example .env
# Add your keys to .env
🛠️ Tech Stack & Architecture

Tech Stack

Layer Technology Purpose
SDK athena Python package Core search, reranking, memory
Reasoning Claude Opus 4.6 (primary) Main reasoning engine
IDE Antigravity (supports Claude, Gemini, GPT) Agentic development environment
Local Models Ollama, LM Studio (any model) Model-agnostic — memory is in files, not the model
Embeddings text-embedding-004 (768-dim) Google embedding model
GraphRAG NetworkX + Leiden + ChromaDB Knowledge graph ⚠️ ~$50 API
Memory Supabase + pgvector or local (ChromaDB) Vector database
Knowledge Store Markdown files (git-versioned) Human-readable, locally owned
Reasoning Engine Graph of Thoughts (GoT) DAG-based planning with aggregation/synthesis
Query Routing CognitiveRouter Adaptive latency based on query complexity
Budget Enforcement BudgetGatekeeper Real-time token/tool/cost limit enforcement

The Core Loop

┌─────────────────────────────────────────────────────────────────────────┐
│                                                                         │
│   (1) /start ──► Retrieve Context ──► (2) Work ──► (3) /end             │
│       ▲                                                    │            │
│       │                                                    ▼            │
│       └───── (5) Next Session ◄── Embed ◄── (4) Extract & Store        │
│                                                                         │
└─────────────────────────────────────────────────────────────────────────┘

Think of it like Git, but for conversations. Each session builds on the last. Important decisions get captured, indexed, and recoverable.

What Athena Does

Feature How It Works
/start boot Loads identity + retrieves relevant context from long-term memory
/end commit Summarizes session, extracts decisions, saves to knowledge store
Hybrid search Fuses Canonical + GraphRAG + Tags + Vectors + Filenames via RRF
MCP Server 9 tools via Model Context Protocol — any MCP client can connect
Permissioning 4 capability levels + 3 sensitivity tiers + Secret Mode (demo/external)
Evaluator Gate 50-query regression suite: MRR@5 = 0.44, Hit@5 = 52%
Cross-encoder reranking Refines top results with sentence-transformers
Protocol library 330 unique protocols (75 curated in starter pack)

Repository Structure

Athena-Public/
├── src/athena/           # SDK package (pip installable)
│   ├── core/             #    Config, models, governance, permissions
│   ├── tools/            #    Search, reranker, latency
│   ├── memory/           #    Vector DB interface
│   └── mcp_server.py     #    MCP Tool Server (9 tools, 2 resources)
├── examples/
│   ├── quickstart/       # Runnable demos
│   ├── scripts/          # Automation scripts
│   ├── protocols/        # Thinking patterns (starter pack included)
│   ├── workflows/        # Slash commands
│   └── templates/        # Starter templates
├── docs/                 # Deep documentation
├── community/            # Contributing, roadmap
├── pyproject.toml        # Modern packaging
└── .env.example          # Environment template
📖 Key Concepts & Workflows

Key Concepts

Example Workflows

Command Description
/start Boot system, load identity
/end Close session, commit to memory
/think Deep reasoning mode
/ultrathink Maximum depth analysis
/brief Pre-prompt clarification protocol
/refactor Workspace optimization
/research Multi-source web research
/plan Structured planning with pre-mortem
/save Mid-session checkpoint
/vibe Ship at 70%, iterate fast

👉 Full Workflow Documentation — Detailed guide to all 12 workflows

See .agent/workflows/ for source files.

📋 Changelog

February 2026

  • v8.5.0 (Feb 12 2026): Phase 1 Complete — MCP Tool Server (9 tools, 2 resources), Permissioning Layer (4 levels + secret mode), Search MRR +105% (0.21→0.44), Evaluator Gate (50 queries). SDK v2.0.0.
  • v8.3.1 (Feb 11 2026): Viral Validation Release — 654K+ Reddit views, 1,660+ upvotes, 5,300+ shares. #1 All-Time r/ChatGPT, #2 All-Time r/GeminiAI. Three-Phase Token Budget: robustness at /start+/end, Adaptive Latency in-between.
  • v8.2.1 (Feb 9 2026): Metrics Sync — Fixed batch_audit.py automation, linked orphan files, reconciled tech debt, 8,079 tags indexed

👉 Full Changelog — Complete version history from v1.0.0 (Dec 2025)


For the full documentation, case studies, and deep dives, see docs/.

About

The Linux OS for AI Agents — A local OS that gives any LLM persistent memory, autonomy, and time-awareness. Own the state. Rent the intelligence.

Resources

License

Code of conduct

Contributing

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages

  • Python 94.0%
  • PLpgSQL 4.8%
  • Shell 1.2%