veRL is a flexible and efficient RL framework for LLMs.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
DeepSpeed version of NVIDIA's Megatron-LM that adds additional support for several features such as MoE model training, Curriculum Learning, 3D Parallelism, and others.
A Native-PyTorch Library for LLM Fine-tuning.
A native PyTorch Library for large model training.
Generative AI framework built for researchers and PyTorch developers working on Large Language Models (LLMs), Multimodal Models (MMs), Automatic Speech Recognition (ASR), Text to Speech (TTS), and Computer Vision (CV) domains.
Ongoing research training transformer models at scale.
Making large AI models cheaper, faster, and more accessible.
Efficient Training for Big Models.
Mesh TensorFlow: Model Parallelism Made Easier.
A simple, performant and scalable Jax LLM!
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
A library for accelerating Transformer model training on NVIDIA GPUs.
veRL is a flexible and efficient RL framework for LLMs.