data-parallelism topic

List data-parallelism repositories

DeepSpeed

33.0k
Stars
3.9k
Forks
290
Watchers

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

weave

529
Stars
22
Forks
Watchers

A state-of-the-art multithreading runtime: message-passing based, fast, scalable, ultra-low overhead

dist-keras

623
Stars
170
Forks
Watchers

Distributed Deep Learning, with a focus on distributed training, using Keras and Apache Spark.

ColossalAI

38.0k
Stars
4.3k
Forks
376
Watchers

Making large AI models cheaper, faster and more accessible

terngrad

181
Stars
48
Forks
Watchers

Ternary Gradients to Reduce Communication in Distributed Deep Learning (TensorFlow)

dkeras

192
Stars
13
Forks
Watchers

Distributed Keras Engine, Make Keras faster with only one line of code.

PaddleFleetX

425
Stars
157
Forks
Watchers

飞桨大模型开发套件,提供大语言模型、跨模态大模型、生物计算大模型等领域的全流程开发工具链。

libai

376
Stars
55
Forks
Watchers

LiBai(李白): A Toolbox for Large-Scale Distributed Parallel Training

EasyParallelLibrary

252
Stars
49
Forks
Watchers

Easy Parallel Library (EPL) is a general and efficient deep learning framework for distributed model training.

orkhon

145
Stars
5
Forks
Watchers

Orkhon: ML Inference Framework and Server Runtime