tner icon indicating copy to clipboard operation
tner copied to clipboard

Fine Tuning with a new model

Open tednaseri opened this issue 2 years ago • 1 comments

Hi @asahi417 ,

I am trying to fine-tune using different pre-trainedf distil-bert models. If the number of labels is not matched with the one Tner expects, I face with an error as:

RuntimeError: Error(s) in loading state_dict for DistilBertForTokenClassification:
	size mismatch for classifier.weight: copying a param with shape torch.Size([9, 768]) from checkpoint, the shape in current model is torch.Size([15, 768]).
	size mismatch for classifier.bias: copying a param with shape torch.Size([9]) from checkpoint, the shape in current model is torch.Size([15]).
	You may consider adding `ignore_mismatched_sizes=True` in the model `from_pretrained` method.

The suggested solution is using ignore_mismatched_sizes=True when loading model, like: loading --> from_pretrained(path, num_labels, ignore_mismatched_sizes=True) What do you think about it?

Thank you.

tednaseri avatar Nov 09 '22 03:11 tednaseri