Adrien Merat
Adrien Merat
https://github.com/apache/airflow/pull/38969 Is patched from my point of view. Thanks Idan Tepper!
> I'm not certain this is the problem so it would be good to validate it. But fusing can cause precision issues. In low precision: `c = a + b`...
Here's my loss when running the original script with 13k training samples, 3k validation samples (random split) over 1000 iterations with learning rate of 1e-6, rank of 16 and scale...
I'm getting the same `unknown data type: U8` errors from ollama when creating an assistant from Llama-3.1 after LoRA quantization and fine-tuning with custom data. If I try to convert...
> @Timelessprod what framework did you use to create the model? Can you provide the `ollama create` line and the Modelfile, and is it possible to get access to the...
@pdevine In my case I am limited by resources that's why I need to rely on quantization and BitsAndBytes seemed to be quite standard (it's the default quantization used when...
@pdevine What I meant is that BitsAndBytes is encoding my fine-tuned model with U8 tensors which are not supported by Ollama. Ollama only handle F16, FB16 and F32 dtypes, cf....
Okay I understand thank you. Indeed with Unsloth I have no problem with the dtype used for tensors.