data-parallelism topic
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
weave
A state-of-the-art multithreading runtime: message-passing based, fast, scalable, ultra-low overhead
dist-keras
Distributed Deep Learning, with a focus on distributed training, using Keras and Apache Spark.
ColossalAI
Making large AI models cheaper, faster and more accessible
terngrad
Ternary Gradients to Reduce Communication in Distributed Deep Learning (TensorFlow)
dkeras
Distributed Keras Engine, Make Keras faster with only one line of code.
PaddleFleetX
飞桨大模型开发套件,提供大语言模型、跨模态大模型、生物计算大模型等领域的全流程开发工具链。
libai
LiBai(李白): A Toolbox for Large-Scale Distributed Parallel Training
EasyParallelLibrary
Easy Parallel Library (EPL) is a general and efficient deep learning framework for distributed model training.
orkhon
Orkhon: ML Inference Framework and Server Runtime