Get started with language model post-training using Training Hub
Introducing Training Hub: An open source, algorithm-centered library for LLM training.
A collective of researchers and engineers from Red Hat & IBM building LLM toolkits you can use today.
Inference-time scaling for LLMs.
Synthetic data generation pipelines
Post training algorithms for LLMs
Asynchronous GRPO for scalable reinforcement learning.
A method for skipping redundant attention blocks in language models
Efficient training library for large language models up to 70B parameters on a single node.
Adaptive SVD-based continual learning method for LLMs.
Inference-time scaling with particle filtering.
State-of-the-art reward models for preference data generation and acceptance criteria.
KV cache quantization for scaling inference time
Efficient messages-format SFT library for language models
Introducing Training Hub: An open source, algorithm-centered library for LLM training.
Post-training adapts language models for specific, safe, and practical uses. This overview highlights key methods and the open source training_hub library.
Customize reasoning models with synthetic data generation for enterprise deployment. Learn techniques from Red Hat's AI Innovation Team.
👤 Speaker: Young Jin Park
👤 Speaker: Mustafa Eyceoz
👤 Speaker: Shivchander Sudalairaj & Abhishek Bhandwaldar