Jiahui Chen
Results
1
comments of
Jiahui Chen
`--model_max_length 32768 \` Needs to be set for any SigLIP run or else the loss is constant and grad_norm is 0.0 for any fine-tuning run, not just with LORA finetuning