unsloth
unsloth copied to clipboard
[WIP] add support for mixtral
Mixtral WIP
Fantastic and fabulous work @tohrnii!!! Super appreciate it! I will take a look later today!
Any update about this pull request?
@kaykyr @danielhanchen @tohrnii You guys open to some collaboration on this? I think I just my Phi2 implementation done (big touch wood) so I'm happy to take a look
Apologies, I got stuck on something else. I'd love to collaborate @cm2435. If however you are close to completing the implementation, I'm happy to close this PR in favor of yours.
Hey - thanks again on the PR @tohrnii and super appreciate it again :) Ye more than happy to make this happen and collab with you all @kaykyr @cm2435 - I was just a bit bogged down recently on chat templates and making a UI - I will be much more free next week, then we can make Mixtral happen :)
For sure! I am trying to fine tune a MoE pretrained if I had progress I will create pull requests guys. I also able to offer my small server (2x RTX 3090 with NVLink + i9 11900HK + 64GB DDR4) for collaborators who wanna run tests with multi-gpu.
@kaykyr Oh thanks for the kind offer!!! I'll take up for that offer later in the month :)
@kaykyr funny you mention that- I've got almost the exact same setup! I'm going to be very sad when they deprecate the SLI bridge as cuda supported hardware
Great work. Is there any estimate about when this will be merged?