simple-llm-finetuner icon indicating copy to clipboard operation
simple-llm-finetuner copied to clipboard

Simple UI for LLM Model Finetuning

Results 40 simple-llm-finetuner issues
Sort by recently updated
recently updated
newest added

**In trainer.py, ignore the last token is not suitable for all situations.** ```python def tokenize_sample(self, item, max_seq_length, add_eos_token=True): assert self.tokenizer is not None result = self.tokenizer( item["text"], truncation=True, max_length=max_seq_length, padding="max_length",...

This pull request introduces the `--use_cpu` option to the main script, allowing users to run training and inference on the CPU. The `--use_cpu` option can be specified when running the...

Running inference gives the following warning: ``` Loading tokenizer... The tokenizer class you load from this checkpoint is not the same type as the class this function is called from....

question

Hi there! I would like to know how we can run this solution in multi gpu environment for bigger models. Thank you

While I run the train in colab, this error is shown - ``` Something went wrong Connection errored out. ``` How can I solve this?

colab

Training on T4: torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 44.00 MiB (GPU 0; 14.56 GiB total capacity; 13.25 GiB already allocated; 10.44 MiB free; 13.83 GiB reserved in...

You did an amazing job. But most people wants a cli version to quickly get started and play around. Here it is. Cheers.

I have added progressive output to the inference tab by converting the `generate` function in app.py to a Python generator and producing tokens one at at time until either the...