deepspeed topic
finetune-gpt2xl
Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpeed
gpt-neo-fine-tuning-example
Fine-Tune EleutherAI GPT-Neo And GPT-J-6B To Generate Netflix Movie Descriptions Using Hugginface And DeepSpeed
revlib
Simple and efficient RevNet-Library for PyTorch with XLA and DeepSpeed support and parameter offload
pytorch-gpt-x
Implementation of autoregressive language model using improved Transformer and DeepSpeed pipeline parallelism.
ai-msgbot
Training & Implementation of chatbots leveraging GPT-like architecture with the aitextgen package to enable dynamic conversations.
iam-crnn-ctc-recognition
IAM Dataset Handwriting Recognition Using CRNN, CTC Loss, DeepSpeech Beam Search, And KenLM Scorer
glake
GLake: optimizing GPU memory management and IO transmission.
LLaMA-Cult-and-More
Large Language Models for All, 🦙 Cult and More, Stay in touch !
LearnDeepSpeed
DeepSpeed教程 & 示例注释 & å¦ä¹ 笔记 (大模型高效è®ç»ƒï¼‰
CoLLiE
Collaborative Training of Large Language Models in an Efficient Way