LLamaSharp icon indicating copy to clipboard operation
LLamaSharp copied to clipboard

Is it possible to train or fine tune a model with LLamaSharp?

Open ugurkaancalim opened this issue 1 year ago • 12 comments

Hi,

How can i fine tune or train a llama model with LLamaSharp? I couldn't find a documentation about fine-tuning, training a model or using LLamaSharp Library for any other purposes.

Thank you.

ugurkaancalim avatar Jun 09 '23 14:06 ugurkaancalim

Hi, LLamaSharp cannot train or fine-tune a llama model till now. llama.cpp added such an feature about 20 days ago, however it's not stable and only a demo at the begginning, as you can see in https://github.com/ggerganov/llama.cpp/blob/master/examples/baby-llama/baby-llama.cpp. We'll include this feature in the next 2-3 minor version to follow up llama.cpp.

As for other usings, the integration with semantic-kernel and BotSharp are under work now, as you can see in semantic kernel integration and BotSharp. Through these two users can apply LLamaSharp with more scenarios. For example semantic kernel provides rich features to handle LLM and chat AI; BotSharp makes it easy to deploy chat bot and is compatible with Chat Bot UI.

AsakusaRinne avatar Jun 09 '23 14:06 AsakusaRinne

Hi,

Nice to hear that LLamaSharp will be trainable with C#.

Thank you for the answer.

ugurkaancalim avatar Jun 09 '23 14:06 ugurkaancalim

Hi,

Is there any news about this feature.

ugurkaancalim avatar Jun 21 '23 14:06 ugurkaancalim

Hi,

Is there any news about this feature.

Not yet, I'll do it after completing openai style APIs and integrations with semantic-kernel under LLamaSharp v0.4.x

AsakusaRinne avatar Jun 22 '23 02:06 AsakusaRinne

Hi,

Is there any news about this feature.

Not yet, I'll do it after completing openai style APIs and integrations with semantic-kernel under LLamaSharp v0.4.x

Thank you

ugurkaancalim avatar Jun 23 '23 08:06 ugurkaancalim

This is very intriguing, how will this function work? Will it be possible to insert more sets of questions and answers into an existing model?

Hi, Is there any news about this feature.

Not yet, I'll do it after completing openai style APIs and integrations with semantic-kernel under LLamaSharp v0.4.x

yakovw avatar Jun 23 '23 08:06 yakovw

@AsakusaRinne do you have any news about this feature? It's possible? Thanks!

rachkoud avatar Aug 29 '23 18:08 rachkoud

Happy to see this being worked on, Thanks!

jihadkhawaja avatar Sep 20 '23 16:09 jihadkhawaja

Me too, please let me know if you need any help with implementing this feature

KSemenenko avatar Nov 09 '23 21:11 KSemenenko

Me too, please let me know if you need any help with implementing this feature

Hi all, I'm sorry for delaying this feature for such a long time. The only problem for me is that my time is limited. I've investigated the implementation in llama.cpp two months ago. However I'm not sure about how many changes after that.

Though I'm not able to support the whole feature as good as the inference now, I could have a try to support part of it first. Could you please tell me what you need to start a fine-tune? (for example, data-format, fine-tune process control, etc...)

AsakusaRinne avatar Nov 10 '23 18:11 AsakusaRinne

Hi @AsakusaRinne ,

I apologize, I have just saw your post. I need data-format and some tips to make data useful for LLama.

Thank you

ugurkaancalim avatar Nov 24 '23 17:11 ugurkaancalim

for me the same, some examples, data format, as you mention =)

KSemenenko avatar Nov 24 '23 19:11 KSemenenko