lcqlalala
Results
2
comments of
lcqlalala
Because we added a new tensor with no corresponding weight in the original llama checkpoint, this new tensor needs to be initialized randomly during training. When we load the model...
The tensor we are adding is: self.prompt = torch.nn.parameter.Parameter(torch.randn(self.embed_dim), requires_grad=True) has been initialized. ValueError: prompt is on the meta device, we need a value to put in on 0.