Kartikay Khandelwal
Kartikay Khandelwal
Thanks for the question! We will have more detailed communication around this, but a quick note here. MMF currently supports text + image understanding tasks with some initial support for...
@rohan-varma just BS=1 with the default config.
@joecummings agreed, but with the caveat that we can't have sizable gaps on perf wrt competition. Benchmarking this is a milestone for the MVP.
@walidbet18 thanks for opening this issue! Unfortunately, we don't currently support fine-tuning GGUF models. Is the model you mentioned available in native pytorch format? I'm guessing this is a specific...
Sounds good! Just to confirm, you're looking to fine-tune the mistral 7B model not specifically ``` dolphin-2.2.1-mistral-7b.Q2_K.gguf```, is that right? Or do you care about ``` dolphin-2.2.1-mistral-7b.Q2_K.gguf``` specifically?
Ah ok, then you should be able to use the commands from the README and the tutorials directly, just replace the configs with the one in the mistral folder. So...
@joecummings we probably need to address this as part of ```tune download``` to make sure we arent' downloading to ```tmp```. I can look at this on the checkpointing side as...
Tracking this in #691
Ok yeh, I think I read this as addressing the ```/tmp/artifacts``` issue but that shouldn't be a problem anymore. For downloading the models and files, I think we can find...
Why is what hub does a good idea? Models etc that we're downloading are all from HF, why do we need to do what torch hub does for this? IIUC,...