starcoder
starcoder copied to clipboard
Home of StarCoder: fine-tuning & inference!
Hi friends, I was trying to test the finetune/finetune.py script. It seems that state.best_model_checkpoint always return None leading to a failure at the end of the program. Is it that...
hello after loading the model i asked it what are you able to generate and it responded with a question mark then I asked what project we were working on...
I am exploring the possibility of using StarCoder to generate embeddings for code tokens and would like to know if this is feasible with the current implementation. ### Questions: 1....
When aiming to fine-tune starcoder or octocoder on a custom dataset for integration with an IDE, would it be more appropriate to process the data in a question & answer...
To fix the `unrecognized arguments` problem, when running finetune.py from `torch.distributed.launch`. the argument `local_rank` needs to be changed to `local-rank`. launch command: ```shell python -m torch.distributed.launch --nproc_per_node=2 finetune.py --model_path xxx...
### My System Info peft==0.4.0 accelerate==0.18.0 transformers==4.28.0 py310 ### Reproduction After training, I merge the peft weights with base model using: ``` model_ft = PeftModel.from_pretrained( AutoModelForCausalLM.from_pretrained( base_model_path, return_dict=True, torch_dtype='auto', use_cache=True,...
Even with a NVIDIA A100 80 GB GPU, I am not able to fine tune the model on full sequence length of 8192. I was not able to fine tune...
The HuggingFaceH4/oasst1_en dataset contains "train_idf" and "test_idf" instead of "train" and "test"
Exception in thread Thread-7: Traceback (most recent call last): File "/data/starCoder/software/conda/envs/torch/lib/python3.8/threading.py", line 932, in _bootstrap_inner self.run() File "/data/starCoder/software/conda/envs/torch/lib/python3.8/threading.py", line 870, in run self._target(*self._args, **self._kwargs) File "/data/starCoder/software/conda/envs/torch/lib/python3.8/site-packages/torch/utils/_contextlib.py", line 115, in decorate_context...
"I don't want to use 8-bit training. I hope to use fp16 training. After commenting out these two lines, there was an error. How should I modify it? In addition,...