MCPHub LabRegistryOpenPipe/ART
OpenPipe

OpenPipe/ART

Built by OpenPipe โ€ข 9,079 stars

What is OpenPipe/ART?

Agent Reinforcement Trainer: train multi-step agents for real-world tasks using GRPO. Give your agents on-the-job training. Reinforcement learning for Qwen3.5, GPT-OSS, Llama, and more!

How to use OpenPipe/ART?

1. Install a compatible MCP client (like Claude Desktop). 2. Open your configuration settings. 3. Add OpenPipe/ART using the following command: npx @modelcontextprotocol/openpipe-art 4. Restart the client and verify the new tools are active.
๐Ÿ›ก๏ธ Scoped (Restricted)
npx @modelcontextprotocol/openpipe-art --scope restricted
๐Ÿ”“ Unrestricted Access
npx @modelcontextprotocol/openpipe-art

Key Features

Native MCP Protocol Support
Real-time Tool Activation & Execution
Verified High-performance Implementation
Secure Resource & Context Handling

Optimized Use Cases

Extending AI models with custom local capabilities
Automating system workflows via natural language
Connecting external data sources to LLM context windows

OpenPipe/ART FAQ

Q

Is OpenPipe/ART safe?

Yes, OpenPipe/ART follows the standardized Model Context Protocol security patterns and only executes tools with explicit user-granted permissions.

Q

Is OpenPipe/ART up to date?

OpenPipe/ART is currently active in the registry with 9,079 stars on GitHub, indicating its reliability and community support.

Q

Are there any limits for OpenPipe/ART?

Usage limits depend on the specific implementation of the MCP server and your system resources. Refer to the official documentation below for technical details.

Official Documentation

View on GitHub
<div align="center">

<a href="https://art.openpipe.ai"><picture> <img alt="ART logo" src="https://github.com/openpipe/art/raw/main/assets/ART_logo.png" width="160px"> </picture></a>

<p align="center"> <h1>Agent Reinforcement Trainer</h1> </p> <p> Train multi-step agents for real-world tasks using GRPO. </p>

PRs-Welcome PyPI version Train Agent

Join Discord Documentation

</div>

๐Ÿš€ W&B Training: Serverless RL

W&B Training (Serverless RL) is the first publicly available service for flexibly training models with reinforcement learning. It manages your training and inference infrastructure automatically, letting you focus on defining your data, environment and reward functionโ€”leading to faster feedback cycles, lower costs, and far less DevOps.

โœจ Key Benefits:

  • 40% lower cost - Multiplexing on shared production-grade inference cluster
  • 28% faster training - Scale to 2000+ concurrent requests across many GPUs
  • Zero infra headaches - Fully managed infrastructure that stays healthy
  • Instant deployment - Every checkpoint instantly available via W&B Inference
# Before: Hours of GPU setup and infra management
# RuntimeError: CUDA error: out of memory ๐Ÿ˜ข

# After: Serverless RL with instant feedback
from art.serverless.backend import ServerlessBackend

model = art.TrainableModel(
  project="voice-agent",
  name="agent-001",
  base_model="OpenPipe/Qwen3-14B-Instruct"
)

backend = ServerlessBackend(
    api_key="your_wandb_api_key"
)
model.register(backend)
# Edit and iterate in minutes, not hours!

๐Ÿ“– Learn more about W&B Training โ†’

ART Overview

ART is an open-source RL framework that improves agent reliability by allowing LLMs to learn from experience. ART provides an ergonomic harness for integrating GRPO into any python application. For a quick hands-on introduction, run one of the notebooks below. When you're ready to learn more, check out the docs.

๐Ÿ“’ Notebooks

Agent TaskExample NotebookDescriptionComparative Performance
ARTโ€ขE [Serverless]๐Ÿ‹๏ธ Train agentQwen3 14B learns to search emails using RULER<img src="https://github.com/openpipe/art/raw/main/assets/benchmarks/email_agent/accuracy-training-progress.svg" height="72"> benchmarks
2048 [Serverless]๐Ÿ‹๏ธ Train agentQwen3 14B learns to play 2048<img src="https://github.com/openpipe/art/raw/main/assets/benchmarks/2048/accuracy-training-progress.svg" height="72"> benchmarks
ARTโ€ขE LangGraph๐Ÿ‹๏ธ Train agentQwen 2.5 7B learns to search emails using LangGraph[Link coming soon]
MCPโ€ขRL๐Ÿ‹๏ธ Train agentQwen 2.5 3B masters the NWS MCP server[Link coming soon]
Temporal Clue๐Ÿ‹๏ธ Train agentQwen 2.5 7B learns to solve Temporal Clue[Link coming soon]
Tic Tac Toe๐Ÿ‹๏ธ Train agentQwen 2.5 3B learns to play Tic Tac Toe<img src="https://github.com/openpipe/art/raw/main/assets/benchmarks/tic-tac-toe-local/accuracy-training-progress.svg" height="72"> benchmarks
Codenames๐Ÿ‹๏ธ Train agentQwen 2.5 3B learns to play Codenames<img src="https://github.com/openpipe/art/raw/main/assets/benchmarks/codenames/win_rate_over_time.png" height="72"> benchmarks
AutoRL [RULER]๐Ÿ‹๏ธ Train agentTrain Qwen 2.5 7B to master any task[Link coming soon]
Distillation (SFT)๐Ÿ‹๏ธ Train modelDistill text-to-SQL from Qwen 3 235B to Qwen 3 30B[Link coming soon]
Summarizer (SFT + RL)๐Ÿ‹๏ธ Train modelTrain a document summarizer with SFT warmup then RL[Link coming soon]
SFT from a dataset๐Ÿ‹๏ธ Train modelFine-tune Qwen 3 30B on text-to-SQL from a dataset[Link coming soon]

๐Ÿ“ฐ ART News

Explore our latest research and updates on building SOTA agents.

๐Ÿ“– See all blog posts โ†’

Why ART?

  • ART provides convenient wrappers for introducing RL training into existing applications. We abstract the training server into a modular service that your code doesn't need to interface with.
  • Train from anywhere. Run the ART client on your laptop and let the ART server kick off an ephemeral GPU-enabled environment, or run on a local GPU.
  • Integrations with hosted platforms like W&B, Langfuse, and OpenPipe provide flexible observability and simplify debugging.
  • ART is customizable with intelligent defaults. You can configure training parameters and inference engine configurations to meet specific needs, or take advantage of the defaults, which have been optimized for training efficiency and stability.

Installation

ART agents can be trained from any client machine that runs python. To add to an existing project, run this command:

pip install openpipe-art

๐Ÿค– ARTโ€ขE Agent

Curious about how to use ART for a real-world task? Check out the ARTโ€ขE Agent blog post, where we detail how we trained Qwen 2.5 14B to beat o3 at email retrieval!

<img src="https://github.com/openpipe/art/raw/main/assets/ART_E_graphs.png" width="700">

๐Ÿ” Training Loop Overview

ART's functionality is divided into a client and a server. The OpenAI-compatible client is responsible for interfacing between ART and your codebase. Using the client, you can pass messages and get completions from your LLM as it improves. The server runs independently on any machine with a GPU. It abstracts away the complexity of the inference and training portions of the RL loop while allowing for some custom configuration. An outline of the training loop is shown below:

  1. Inference

    1. Your code uses the ART client to perform an agentic workflow (usually executing several rollouts in parallel to gather data faster).
    2. Completion requests are routed to the ART server, which runs the model's latest LoRA in vLLM.
    3. As the agent executes, each system, user, and assistant message is stored in a Trajectory.
    4. When a rollout finishes, your code assigns a reward to its Trajectory, indicating the performance of the LLM.
  2. Training

    1. When each rollout has finished, Trajectories are grouped and sent to the server. Inference is blocked while training executes.
    2. The server trains your model using GRPO, initializing from the latest checkpoint (or an empty LoRA on the first iteration).
    3. The server saves the newly trained LoRA to a local directory and loads it into vLLM.
    4. Inference is unblocked and the loop resumes at step 1.

This training loop runs until a specified number of inference and training iterations have completed.

๐Ÿงฉ Supported Models

ART should work with most vLLM/HuggingFace-transformers compatible causal language models, or at least the ones supported by Unsloth. Gemma 3 does not appear to be supported for the time being. If any other model isn't working for you, please let us know on Discord or open an issue on GitHub!

๐Ÿค Contributing

ART is in active development, and contributions are most welcome! Please see the CONTRIBUTING.md file for more information.

๐Ÿ“– Citation

@misc{hilton2025art,
  author = {Brad Hilton and Kyle Corbitt and David Corbitt and Saumya Gandhi and Angky William and Bohdan Kovalevskyi and Andie Jones},
  title = {ART: Agent Reinforcement Trainer},
  year = {2025},
  publisher = {GitHub},
  journal = {GitHub repository},
  howpublished = {\url{https://github.com/openpipe/art}}
}

โš–๏ธ License

This repository's source code is available under the Apache-2.0 License.

๐Ÿ™ Credits

ART stands on the shoulders of giants. While we owe many of the ideas and early experiments that led to ART's development to the open source RL community at large, we're especially grateful to the authors of the following projects:

Finally, thank you to our partners who've helped us test ART in the wild! We're excited to see what you all build with it.

Global Ranking

8.5
Trust ScoreMCPHub Index

Based on codebase health & activity.

Manual Config

{ "mcpServers": { "openpipe-art": { "command": "npx", "args": ["openpipe-art"] } } }