Mesh TensorFlow: Model Parallelism Made Easier.
DeepSpeed version of NVIDIA's Megatron-LM that adds additional support for several features such as MoE model training, Curriculum Learning, 3D Parallelism, and others.
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
Generative AI framework built for researchers and PyTorch developers working on Large Language Models (LLMs), Multimodal Models (MMs), Automatic Speech Recognition (ASR), Text to Speech (TTS), and Computer Vision (CV) domains.
A Native-PyTorch Library for LLM Fine-tuning.
Efficient Training for Big Models.
veRL is a flexible and efficient RL framework for LLMs.
Your email address will not be published. Required fields are marked *
Comment *
Name *
Email *
Website
Captcha: 11 - 12 = ?*
Save my name, email, and website in this browser for the next time I comment.
DeepSpeed version of NVIDIA's Megatron-LM that adds additional support for several features such as MoE model training, Curriculum Learning, 3D Parallelism, and others.