megablocks
megablocks copied to clipboard
To setup CI/CD for MegaBlocks - [x] testing #127 - [x] fix setup.py #128 - [ ] Format + lint #131 - [ ] Type checking - [ ] Cut...
the CAPACITY_FACTOR is set improperly. 
when I run /exp/dmoe/dmoe_46m_8gpu.sh,I encountered the following Error during evaluation.  the func **save_load_balancing_loss** is bypassed during evaluation.
Routing
Is the router implemented the noisy top k routing suggested by the [OUTRAGEOUSLY LARGE NEURAL NETWORKS: THE SPARSELY-GATED MIXTURE-OF-EXPERTS LAYER](https://arxiv.org/pdf/1701.06538) paper? In the router code you seem to apply the...
When the input tensor is not on device 0, `histogram` causes an illegal memory access which prevents `indices_and_bins` from being computed correctly on a model & inputs which aren't on...
I use megablocks to implement a fine-granded moe, the ffn_hidden_size is divisible by 64, but is not divisible by 128, can we change it to 64? Thanks a lot
I am trying to setup and use megablocks to train MoE models, but I see the following error: ``` Traceback (most recent call last): File "/n/holyscratch01/dam_lab/brachit/moes/megablocks/third_party/Megatron-LM/pretrain_gpt.py", line 8, in from...
hello, i have tried to use megablocks in V100 + pytorch2.4.0+cu121, but get error with "cannot support bf16". If i use megablocks in fp32, i get error "group gemm must...
Thanks for your work. i check the code and find the mlp only have w1 and w2. Does the sparse mlp support the bias? thanks a lot! i want to...
Do you know what (if anything) stands in the way of using megablocks with torch.compile?