DeepSeek-Coder-V2
DeepSeek-Coder-V2 copied to clipboard
add fine-tuning code with lora support
Closes #35 #41 #58
@guoday
I need this to fine-tune
I remember you telling me about Deepseek in May, props to you @Muhtasham 🏃♂️
@Muhtasham Should the prompt being built in build_instruction_prompt not match the example in the README.md:
<|begin▁of▁sentence|>User: {user_message_1}
Assistant: {assistant_message_1}<|end▁of▁sentence|>User: {user_message_2}
Assistant:
Apologies if I misunderstand.
@adamreed90 Yes this PR is specifically for instruction fine-tuning, so the prompt format in build_instruction_prompt is intentionally different from the chat-based format:
As I pointed in the new README.MD, for training data preparation, please follow the Sample Dataset Format.
If you’re bringing a dataset in a different format (such as chat-based), it would require modification.