veRL is a flexible and efficient RL framework for LLMs.
Ongoing research training transformer models at scale.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Efficient Training for Big Models.
A simple, performant and scalable Jax LLM!
A Native-PyTorch Library for LLM Fine-tuning.
A library for accelerating Transformer model training on NVIDIA GPUs.
Your email address will not be published. Required fields are marked *
Comment *
Name *
Email *
Website
Captcha: 12 + 19 = ?*
Save my name, email, and website in this browser for the next time I comment.
Ongoing research training transformer models at scale.