What is Verifiers?
Verifiers enables you to:- Define custom interaction protocols between models and environments
- Build agents, multi-turn conversations, tool-augmented reasoning, and interactive games
- Create reusable evaluation environments with multi-criteria reward functions
- Train models with the included RL trainer (via
vf-rl) or integrate with other RL frameworks
- First-class OpenAI-compatibility for
ChatCompletionsandCompletions - Extensible multi-turn interactions via
MultiTurnEnv - Native tool calling support with
ToolEnv - Modular reward functions through
Rubricclasses - End-to-end async compatibility with sync support where you want it
- Full-spectrum scaling from CPU evaluations in Jupyter to multi-node GPU RL training
- Environments as Python modules for easy installation, sharing, and reuse
Installation
Basic Installation
For evaluation and API model usage:Training Support
For RL training with the included trainer:main with RL extras:
Latest Development Version
To use the latestmain branch:
Development Setup
For contributing to verifiers:Integration with prime-rl
For large-scale FSDP training, see prime-rl.Integration with Prime Intellect Environments Hub
Coming soon.Documentation
Getting Started
Overview — Core concepts and architecture. Start here if you’re new to Verifiers to understand how environments orchestrate interactions. Environments — Creating custom interaction protocols withMultiTurnEnv, ToolEnv, and basic rubrics.