
Announcing Our Official Hugging Face TRL Integration
Written By:
Kamran Bigdely
Published on
Nov 7, 2025
Today is a big milestone for RapidFire AI: we’re officially integrated into Hugging Face’s TRL documentation as a first-class integration. That means TRL users can now discover, install, and run RapidFire AI as the fastest way to compare many fine-tuning/post-training configurations—without changing their workflow.
Why this matters
Teams don’t have the time (or budget) to train one config after another. Our TRL integration lets you launch many TRL configurations concurrently—even on a single GPU—via adaptive, chunk-based scheduling. In internal benchmarks referenced in the TRL page, this delivers ~16–24× higher experimentation throughput than sequential runs, so you reach better models dramatically faster.
What you get, out of the box
Drop-in TRL wrappers — Use RFSFTConfig, RFDPOConfig, and RFGRPOConfig as near-zero-code replacements for TRL’s SFT/DPO/GRPO configs.
Chunk-based concurrent training — We shard data and cycle configs at chunk boundaries to maximize GPU utilization and enable early, apples-to-apples, comparisons.
Interactive Control Ops (IC Ops) — From the dashboard, Stop, Resume, Clone, and Clone & Warm-Start any run in flight to double-down on winners and pause stragglers—no job restarts required.
Multi-GPU orchestration — The scheduler auto-distributes configs across available GPUs; you focus on models, not plumbing.
MLflow-based dashboard — Real-time metrics, logs, and IC Ops in one place as soon as you start your experiment.
How it works
RapidFire AI slices your dataset into “chunks” and rotates configurations through the GPU at chunk boundaries. You get incremental signal on all configs quickly, while automatic checkpointing keeps training stable. Then, use IC Ops to adapt mid-flight—stop low-performers early and clone promising ones with tweaked hyperparameters (optionally warm-starting from the parent’s weights).
Supported TRL trainers
SFT with
RFSFTConfigDPO with
RFDPOConfigGRPO with
RFGRPOConfig
These are designed as drop-in replacements, so you keep your TRL mental model while gaining concurrency and control.
Get started in minutes
Once running, open the dashboard at http://0.0.0.0:3000, point your code at our TRL-compatible configs, and launch a grid or random search across multiple configurations. You’ll see all runs stream into the dashboard immediately—and you can steer them live with IC Ops.
Where to learn more
We built RapidFire AI so practitioners can iterate faster, learn more from the same GPUs, and ship better LLMs with confidence. Being featured in TRL’s docs validates that mission, and we’re just getting started.
If you try the integration, we’d love to hear how much faster your experimentation loop gets and what you’d like us to add next.

