mixture-of-experts topic

List mixture-of-experts repositories

DeepSpeed

33.2k
Stars
3.9k
Forks
290
Watchers

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

learning-at-home

51
Stars
1
Forks
Watchers

"Towards Crowdsourced Training of Large Neural Networks using Decentralized Mixture-of-Experts" (NeurIPS 2020), original PyTorch implementation

Mixture-Models

45
Stars
17
Forks
Watchers

Hierarchical Mixture of Experts,Mixture Density Neural Network

mtlearn

52
Stars
5
Forks
Watchers

Multi-Task Learning package built with tensorflow 2 (Multi-Gate Mixture of Experts, Cross-Stitch, Ucertainty Weighting)

hivemind

1.8k
Stars
139
Forks
Watchers

Decentralized deep learning in PyTorch. Built to train models on thousands of volunteers across the world.

keras-mmoe

678
Stars
217
Forks
Watchers

A TensorFlow Keras implementation of "Modeling Task Relationships in Multi-task Learning with Multi-gate Mixture-of-Experts" (KDD 2018)

mixture-of-experts

865
Stars
92
Forks
Watchers

PyTorch Re-Implementation of "The Sparsely-Gated Mixture-of-Experts Layer" by Noam Shazeer et al. https://arxiv.org/abs/1701.06538

tutel

670
Stars
81
Forks
Watchers

Tutel MoE: An Optimized Mixture-of-Experts Implementation

mixture-of-experts

556
Stars
45
Forks
Watchers

A Pytorch implementation of Sparsely-Gated Mixture of Experts, for massively increasing the parameter count of language models