TinyLLaVA_Factory
TinyLLaVA_Factory copied to clipboard
template differences?
Are there any differences in the _make_masks function across different LLM models? Don't they all compute loss only for the response part? What causes the variations among them?
Different models use different tokenizers, and when different tokenizers tokenize the text, the corresponding label positions are different.
@jiajunlong Hi. Does llama2, tinyllama and vicuna share the same template? Sorry to bother you.