model-parallelism topic

List model-parallelism repositories

DeepSpeed

33.2k
Stars
3.9k
Forks
290
Watchers

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

ColossalAI

38.1k
Stars
4.3k
Forks
376
Watchers

Making large AI models cheaper, faster and more accessible

PaddleFleetX

425
Stars
158
Forks
Watchers

飞桨大模型开发套件,提供大语言模型、跨模态大模型、生物计算大模型等领域的全流程开发工具链。

libai

377
Stars
55
Forks
Watchers

LiBai(李白): A Toolbox for Large-Scale Distributed Parallel Training

EasyParallelLibrary

252
Stars
49
Forks
Watchers

Easy Parallel Library (EPL) is a general and efficient deep learning framework for distributed model training.

torchgpipe

782
Stars
94
Forks
Watchers

A GPipe implementation in PyTorch

veloce

18
Stars
0
Forks
Watchers

WIP. Veloce is a low-code Ray-based parallelization library that makes machine learning computation novel, efficient, and heterogeneous.

awesome-distributed-ml

166
Stars
18
Forks
Watchers

A curated list of awesome projects and papers for distributed training or inference

redco

41
Stars
3
Forks
Watchers

NAACL '24 (Demo) / MlSys @ NeurIPS '23 - RedCoast: A Lightweight Tool to Automate Distributed Training and Inference

pipegoose

74
Stars
17
Forks
Watchers

Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*