Daniel Han
Daniel Han
@daegonYu Yes that should work (I think) - The continued pretraining notebook does train on the same LoRA adapters twice - https://colab.research.google.com/drive/1tEd1FrOXWMnCU9UIvdYhs61tkxdMuKZu?usp=sharing so it should function (hopefully)
@daegonYu You might be interested in our conversational notebook which masks out the instruction - https://colab.research.google.com/drive/1T5-zKWM_5OD21QHwXHiV9ixTRR7k3iB9?usp=sharing Also see https://github.com/unslothai/unsloth/wiki#train-on-completions--responses-only-do-not-train-on-inputs
@daegonYu Sorry on the delay! Yes they're equivalent EXCEPT if you're doing more than 1 conversation. HF's one does not support it, whilst Unsloth does.
@Candice1995 Apologies on the delay - DPO has a prompt, then 2 other fields - the accepted or rejected answer to the prompt. These fields have varying lengths, and so...
Thanks! Is there a way to actually make `calculate_loan_emi` executable? Maybe via exec / eval?
Apologies just fixed - please do `pip install --force-reinstall --no-deps --no-cache-dir unsloth unsloth_zoo`
Will fix this asap!
Is this single batched inference?
@ziemowit-s Maybe I might have solved it, but unsure with yesterday's patch
Currently not, but we plan to provide that in the future - currently please use HuggingFace