TRL
Hugging Face library for training language models with RLHF, SFT, and DPO
TRL (Transformer Reinforcement Learning) is the standard Hugging Face library for fine-tuning language models. It supports supervised fine-tuning (SFT), reinforcement learning from human feedback (RLHF), direct preference optimization (DPO), and other alignment techniques. Built on top of Transformers and integrates with PEFT for parameter-efficient training.
Pricing: Free
TRL Alternatives
Explore 21 products in the Fine-tuning category. View all TRL alternatives.
Hugging Face
The open-source AI platform with 500K+ models, inference endpoints, and fine-tuning tools
fal
Build the next generation of creativity with fal. Lightning fast inference.
OpenAI
API access to GPT, o-series reasoning, DALL-E, and Whisper models
Amazon Bedrock
Managed API access to foundation models on AWS with built-in fine-tuning and agent tooling
Is your product missing?