Taha Binhuraib
Taha Binhuraib
`weight_quantizer.quantize(transpose(sd[0][prefix + 'self_attention.query_key_value.' + 'weight']))) File "/opt/conda/lib/python3.7/site-packages/deepspeed/module_inject/replace_module.py", line 100, in copy dim=self.in_dim)[self.gpu_index].to(` This is the error I got today while trying int8 inference with bloom.
Thanks @mrwyattii can't wait!
@buswrecker I can run deepspeed mii from the gpu vm but I still can't deploy, I get the same error: `subprocess.CalledProcessError: Command '['az', 'ml', 'workspace', 'show', '--query', 'container_registry']' returned non-zero...
Does the model path mentioned in the readme have to be a huggingface model? If so, what does downloading the binary do?
ValueError: Tokenizer class LLaMATokenizer does not exist or is not currently imported.
Fixed it with this PR: https://github.com/nomic-ai/nomic/pull/132
@vesuppi which GPU are you using? Also, could you share your loss curves? I'm trying to pretrain on a single RTX 4090.
@karan78tp this is with 2 gpus right?
Default arguments in the train.py file. The dataset: openweb text.