Roya
Roya
I don't know if you've ever encountered this issue in your training, the loss increased but all other metrics decreased. It's weird... I don't know if it affects convergence.
Hello!!! I just can't replicate your reslut on snips when using bert + CRF, it's the baseline bert-base-uncased or any tricks when implement training?
I wanna to ask why loss is 0 at random in the same experimental setting when runing finetune_toy_low_resource.sh in ltu_as code or have you ever meet this problem. Do you...
I'd like to ask is there stage training sh scripts for low resource. It seems there is only low resource training sh scripts for finetune toy data. 
I want to ask where can download the whisper model file end with .pt. 
I notice a lot of sh script in the directory train_scripts, I want to ask what's the difference between them. 
There is no code related with training, is this code incomplete or there are no plans to open source the training code
deepspeed报错
 佬我把sh脚本里的模型改成了没有Instruct过的llama,但是运行torchrun的时候报错了,这个要怎么解决呀