
Training Frameworks
GPT-NeoX
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.