llm2vec icon indicating copy to clipboard operation
llm2vec copied to clipboard

can the llm2vec model be reused for chat (post LLM2vec fine-tune for SFT) ?

Open StephennFernandes opened this issue 1 year ago • 2 comments

Hi there, I am building multilingual chat LLMs often in Encoder, EmcDec models like XLM-R, mT6 and parallel bilingual corpus is used for model to learn cross lingual references and generalizes on multilingual dataset by cross lingual generalization.

given that CLM is really bad for such tasks, LLM2vec seems like there is still hope for training Decoder only LLMs on such corsa lingual generalizations.

I just wanted to know if I could post training multilingual LLMs on LLM2vec style training could I use the same LLM for SFT training later on ?

StephennFernandes avatar May 07 '24 09:05 StephennFernandes

I have the same concern,.

louieworth avatar May 07 '24 13:05 louieworth

I did not fully understand your question. All LLM2Vec models and code is public and can be used for further fine-tuning on any dataset/task.

We have not trained LLM2Vec on any cross lingual tasks, but you can definitely go ahead and do it.

Let me know if you have any more concerns or questions

vaibhavad avatar May 07 '24 22:05 vaibhavad

Feel free to re-open if you have any more questions regarding this issue.

vaibhavad avatar May 13 '24 15:05 vaibhavad