Rasmus Toivanen
Rasmus Toivanen
Ok I am now back at trying this. I downloaded the Finnish Fleurs data with the provided download scripts and created the manifest files. Now I am trying to finetune...
Any better example on how to do finetuning? Do you have a test that the text to speech finetuning should work? I tried finetuning with Fleurs data and it threw...
I could take on the task of translating HellaSwag samples for Reasoning. I have some translation credits left for this month in Deepl suscription so that could be used
I am having similar issues:   GGUF files from here: https://huggingface.co/mradermacher/Ahma-3B-Instruct-GGUF/tree/main Original model: https://huggingface.co/Finnish-NLP/Ahma-3B-Instruct
From here you should find conversion script for gemma 2 https://github.com/AI-Hypercomputer/maxtext/issues/1324
Hopefully this get's merged soon. I thought it would be easy to implement some custom_metrics to calculate like mtbench scores at every eval step as previously something like bleu/wer were...
What is your eval_batch_size? If not defined huggingface probably still defaults to 8
I've had also issues. Everything seemed to work fine with unsloth 2025.10.1. Now I am doing similar training with Gemma3 (Same bs, ga, precision etch) but memory is not stable...
I have been able to work on 2025.10.1. I will try the latest release again later
Have you tried reducing rank. That has small impact but still might be worth trying