MentalLLaMA icon indicating copy to clipboard operation
MentalLLaMA copied to clipboard

About training code and scripts

Open wytbwytb opened this issue 1 year ago • 6 comments

It's a nice work. When will you upload the training code and scripts ?

wytbwytb avatar Jan 17 '24 09:01 wytbwytb

We mostly modified the scripts of FastChat (https://github.com/lm-sys/FastChat) for the fine-tuning process. You can look into that.

SteveKGYang avatar Jan 17 '24 10:01 SteveKGYang

Hi @SteveKGYang, great work! Can you please release the code used for training bart-large and T5?

NirmalManoj avatar Feb 06 '24 16:02 NirmalManoj

@SteveKGYang I want to fine-tune with bart-base, but with the same code, processing, etc. that your team used.

NirmalManoj avatar Feb 06 '24 19:02 NirmalManoj

Why not just release it bro

biirving avatar Jun 16 '24 04:06 biirving

Which fine-tuning method you guys used for this ?

  • Unfreezing only last few layers
  • LoRA
  • Attached an extra decoder layer at last ?

Zuhashaik avatar Jul 13 '24 21:07 Zuhashaik

@Zuhashaik We used full fine-tuning, which means all parameters are tuned.

SteveKGYang avatar Jul 14 '24 03:07 SteveKGYang