OpenChatKit
OpenChatKit copied to clipboard
Why instruction tuning calculate whole sentence loss?
I noticed that OIG dataset adds human and bot tag in each sample. In your code, you directly pack samples to max seq length and calculate cross entropy on whole sentence. Will this make the model output human, bot tag and not knowing when to stop? Does only calculate the last bot response loss be more suitable?