Skip to content

AgentR1/Claw-R1

Repository files navigation

Claw-R1: Empowering OpenClaw with
Advanced Agentic RL.

Project Home GitHub Repo stars GitHub forks Docs

Claw-R1 Logo

News

  • [2026.03] 🚧 Claw-R1 Project Init. We are actively updating the framework. Stay tuned for more features and documentation.

Overview

Agentic RL has become the dominant approach for training powerful LLM agents. Meanwhile, General Agents (e.g., OpenClaw, Claude Code, Open Code, etc.) have emerged as game-changing systems that redefine what agents can do. Yet there remains critical gaps:

  • General Agent for Agentic RL: Traditional Agentic RL frameworks typically rely on simple agents like ReAct. General agents (e.g., OpenClaw, Claude Code, Open Code) offer far richer capabilities—but existing RL pipelines were not designed for them.

  • Agentic RL for General Agent: Modern base models have not been fully adapted to thrive inside general agent architectures. We aim to enable models to play a larger, more effective role within these next-generation agents.

Claw-R1 is training framework that bridges this gap. It introduces a Middleware Layer (Gateway Server + DataPool) as the sole bridge between Agent Side and Training Side. Agents—white-box or black-box—access the framework via standard HTTP. This enables three modes: white-box offline, black-box offline, and black-box online service. No framework today adequately supports this paradigm—Claw-R1 is designed to fill that void.

Claw-R1 Framework

Key Features

  • Asynchronous Training & Rollout: Decouples RL training from rollout in the framework, enabling scalable and efficient data collection and model updates.

  • Agent–Training Decoupling: Supports online-service agents where execution and training run independently. Data flows from live user requests into DataPool; the Trainer continuously fetches batches for training—no dataset required.

  • Zero-Code Intrusion: Black-box agents (LangChain, AutoGen, CrewAI, etc.) integrate with zero modification—just point base_url to the Gateway. The framework automatically collects interaction data and trains models.

Get Started

Explore our comprehensive documentation for setup, configuration, and advanced usage:

Contributors

Team Members: Daoyu Wang, Jie Ouyang, Shuo Yu

Supervisors: Qi Liu, Mingyue Cheng

Affiliation: State Key Laboratory of Cognitive Intelligence, University of Science and Technology of China

Acknowledgements

Claw-R1 builds upon Agent-R1. We extend our gratitude to MiniMax Forge for their architectural insights on the Middleware design, and to rLLM for their pioneering work on RL framework design for language agents. We also thank OpenClaw for their remarkable work on personal AI assistants—the modern agent paradigm that inspires our vision. We are grateful to the broader Agentic RL community and all contributors for their support.

Citation

@misc{clawr1-2026,
  title={Claw-R1: Agentic RL for Modern Agents},
  author={Wang, Daoyu and Ouyang, Jie and Yu, Shuo and Cheng, Mingyue and Liu, Qi},
  year={2025},
  howpublished={\url{https://github.com/AgentR1/Claw-R1}},
  note={GitHub repository}
}

About

Claw-R1: Empowering OpenClaw with Advanced Agentic RL.

Topics

Resources

License

Contributing

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages