mixture-of-experts topic

List mixture-of-experts repositories
trafficstars

Generalizable-Mixture-of-Experts

283
Stars
35
Forks
Watchers

GMoE could be the next backbone model for many kinds of generalization task.

pipegoose

77
Stars
17
Forks
Watchers

Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*

soft-mixture-of-experts

60
Stars
4
Forks
Watchers

PyTorch implementation of Soft MoE by Google Brain in "From Sparse to Soft Mixtures of Experts" (https://arxiv.org/pdf/2308.00951.pdf)

llama-moe

862
Stars
46
Forks
Watchers

⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training (EMNLP 2024)

mixture-of-attention

101
Stars
3
Forks
Watchers

Some personal experiments around routing tokens to different autoregressive attention, akin to mixture-of-experts

Neural-Implicit-Dict

31
Stars
1
Forks
Watchers

[ICML 2022] "Neural Implicit Dictionary via Mixture-of-Expert Training" by Peihao Wang, Zhiwen Fan, Tianlong Chen, Zhangyang Wang

MoSE-AUSeg

24
Stars
1
Forks
Watchers

The official code repo for the paper "Mixture of Stochastic Experts for Modeling Aleatoric Uncertainty in Segmentation". (ICLR 2023)

soft-moe

28
Stars
2
Forks
Watchers

PyTorch implementation of "From Sparse to Soft Mixtures of Experts"

Pytorch_mixture-of-experts

27
Stars
1
Forks
Watchers

PyTorch implementation of moe, which stands for mixture of experts

mixtools

17
Stars
4
Forks
Watchers

Tools for Analyzing Finite Mixture Models