verl
Volcano Engine Reinforcement Learning for LLMs
...It brings together supervised fine-tuning, preference modeling, and online RL into one coherent training stack so teams can move from raw data to aligned policies with minimal glue code. The library focuses on scalability and efficiency, offering distributed training loops, mixed precision, and replay/buffering utilities that keep accelerators busy. It ships with reference implementations of popular alignment algorithms and clear examples that make it straightforward to reproduce baselines before customizing. Data pipelines treat human feedback, simulated environments, and synthetic preferences as interchangeable sources, which helps with rapid experimentation. ...