h2o-llmstudio
h2o-llmstudio copied to clipboard
[CODE IMPROVEMENT] Improve mask prompt labels for chained parent data
The setting Mask Prompt Labels allows to fully mask the prompt labels and only calculate the loss on the output.
When chaining conversation data while training, we will have a mix of prompts and answers though and currently this setting only calculates loss on the last output.
We potentially want to calculate the loss of all outputs in the whole chained conversation.