candle
candle copied to clipboard
Implement DeepSeek V3/R1
- [x] Model
- [x] FP8 weight dequantize
- [x] Tensor parallelism
Seems like most of this PR is adding the f8 dtype, perhaps worth separating that out from the deepseek model addition so its easier to review?
this Would be great. And the F8 type is really helpful for other projects.
it looks like the working branch is behind the main branch now. maybe the fp8 work should be pulled out. it looks like a lot of good work by @EricLBuehler
Interesting idea. I'll take a look at doing that once a few other PRs are merged - specifically the Metal MM speedup one.
Is this supposed to be able to support training?