discoelysiumLW
discoelysiumLW
I ran into this error " ValueError: weight is on the meta device, we need a `value` to put in on cpu." while loading llama-7B in 8 bits as well....
If you load the pre-trained model sequentially seems to fix the problem: `model = transformers.AutoModelForCausalLM.from_pretrained( './llama-7b-hf', load_in_8bit=True, device_map='sequential', )`
For right now the model should be load as: AutoModelForCausalLM also the tokenizer: AutoTokenizer