Home / Fine-tuning / TRL

TRL

Hugging Face library for training language models with RLHF, SFT, and DPO

Open Source Free Trial

TRL (Transformer Reinforcement Learning) is the standard Hugging Face library for fine-tuning language models. It supports supervised fine-tuning (SFT), reinforcement learning from human feedback (RLHF), direct preference optimization (DPO), and other alignment techniques. Built on top of Transformers and integrates with PEFT for parameter-efficient training.

Pricing: Free

Screenshot of TRL webpage

Is your product missing? 👀 Add it here →