discoelysiumLW

Results 3 comments of discoelysiumLW

I ran into this error " ValueError: weight is on the meta device, we need a `value` to put in on cpu." while loading llama-7B in 8 bits as well....

If you load the pre-trained model sequentially seems to fix the problem: `model = transformers.AutoModelForCausalLM.from_pretrained( './llama-7b-hf', load_in_8bit=True, device_map='sequential', )`

For right now the model should be load as: AutoModelForCausalLM also the tokenizer: AutoTokenizer