VisualGLM-6B icon indicating copy to clipboard operation
VisualGLM-6B copied to clipboard

utoModel.from_pretrained("../visualglm-6b", trust_remote_code=True).half().cuda() error

Open wavelet2008 opened this issue 1 year ago • 2 comments

(face19) lan@lan:~/sdf/VisualGLM-6B$ python cli_demo_hf.py Explicitly passing a revision is encouraged when loading a model with custom code to ensure no malicious code has been contributed in a newer revision. Traceback (most recent call last): File "/home/lan/sdf/VisualGLM-6B/cli_demo_hf.py", line 7, in tokenizer = AutoTokenizer.from_pretrained("../visualglm-6b", trust_remote_code=True) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/transformers/models/auto/tokenization_auto.py", line 679, in from_pretrained return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1804, in from_pretrained return cls._from_pretrained( File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1958, in _from_pretrained tokenizer = cls(*init_inputs, **init_kwargs) File "/home/lan/.cache/huggingface/modules/transformers_modules/visualglm-6b/tokenization_chatglm.py", line 221, in init self.sp_tokenizer = SPTokenizer(vocab_file, num_image_tokens=num_image_tokens) File "/home/lan/.cache/huggingface/modules/transformers_modules/visualglm-6b/tokenization_chatglm.py", line 64, in init self.text_tokenizer = TextTokenizer(vocab_file) File "/home/lan/.cache/huggingface/modules/transformers_modules/visualglm-6b/tokenization_chatglm.py", line 22, in init self.sp.Load(model_path) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/sentencepiece/init.py", line 905, in Load return self.LoadFromFile(model_file) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/sentencepiece/init.py", line 310, in LoadFromFile return _sentencepiece.SentencePieceProcessor_LoadFromFile(self, arg) RuntimeError: Internal: src/sentencepiece_processor.cc(1101) [model_proto->ParseFromArray(serialized.data(), serialized.size())]

wavelet2008 avatar May 29 '23 09:05 wavelet2008

(face19) lan@lan:~/sdf/VisualGLM-6B$ python cli_demo_hf.py Explicitly passing a revision is encouraged when loading a model with custom code to ensure no malicious code has been contributed in a newer revision. Traceback (most recent call last): File "/home/lan/sdf/VisualGLM-6B/cli_demo_hf.py", line 7, in tokenizer = AutoTokenizer.from_pretrained("../visualglm-6b", trust_remote_code=True) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/transformers/models/auto/tokenization_auto.py", line 679, in from_pretrained return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1804, in from_pretrained return cls._from_pretrained( File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1958, in _from_pretrained tokenizer = cls(*init_inputs, **init_kwargs) File "/home/lan/.cache/huggingface/modules/transformers_modules/visualglm-6b/tokenization_chatglm.py", line 221, in init self.sp_tokenizer = SPTokenizer(vocab_file, num_image_tokens=num_image_tokens) File "/home/lan/.cache/huggingface/modules/transformers_modules/visualglm-6b/tokenization_chatglm.py", line 64, in init self.text_tokenizer = TextTokenizer(vocab_file) File "/home/lan/.cache/huggingface/modules/transformers_modules/visualglm-6b/tokenization_chatglm.py", line 22, in init self.sp.Load(model_path) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/sentencepiece/init.py", line 905, in Load return self.LoadFromFile(model_file) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/sentencepiece/init.py", line 310, in LoadFromFile return _sentencepiece.SentencePieceProcessor_LoadFromFile(self, arg) RuntimeError: Internal: src/sentencepiece_processor.cc(1101) [model_proto->ParseFromArray(serialized.data(), serialized.size())]

@wavelet2008 有解决吗 我也遇到同样的问题

AdJIa avatar Jun 04 '23 07:06 AdJIa

(face19) lan@lan:~/sdf/VisualGLM-6B$ python cli_demo_hf.py Explicitly passing a revision is encouraged when loading a model with custom code to ensure no malicious code has been contributed in a newer revision. Traceback (most recent call last): File "/home/lan/sdf/VisualGLM-6B/cli_demo_hf.py", line 7, in tokenizer = AutoTokenizer.from_pretrained("../visualglm-6b", trust_remote_code=True) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/transformers/models/auto/tokenization_auto.py", line 679, in from_pretrained return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1804, in from_pretrained return cls._from_pretrained( File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1958, in _from_pretrained tokenizer = cls(*init_inputs, **init_kwargs) File "/home/lan/.cache/huggingface/modules/transformers_modules/visualglm-6b/tokenization_chatglm.py", line 221, in init self.sp_tokenizer = SPTokenizer(vocab_file, num_image_tokens=num_image_tokens) File "/home/lan/.cache/huggingface/modules/transformers_modules/visualglm-6b/tokenization_chatglm.py", line 64, in init self.text_tokenizer = TextTokenizer(vocab_file) File "/home/lan/.cache/huggingface/modules/transformers_modules/visualglm-6b/tokenization_chatglm.py", line 22, in init self.sp.Load(model_path) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/sentencepiece/init.py", line 905, in Load return self.LoadFromFile(model_file) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/sentencepiece/init.py", line 310, in LoadFromFile return _sentencepiece.SentencePieceProcessor_LoadFromFile(self, arg) RuntimeError: Internal: src/sentencepiece_processor.cc(1101) [model_proto->ParseFromArray(serialized.data(), serialized.size())]

@wavelet2008 有解决吗 我也遇到同样的问题

参考一下这个:https://github.com/THUDM/ChatGLM-6B/issues/770

shunyuchu avatar Jun 06 '23 07:06 shunyuchu

(face19) lan@lan:~/sdf/VisualGLM-6B$ python cli_demo_hf.py Explicitly passing a revision is encouraged when loading a model with custom code to ensure no malicious code has been contributed in a newer revision. Traceback (most recent call last): File "/home/lan/sdf/VisualGLM-6B/cli_demo_hf.py", line 7, in tokenizer = AutoTokenizer.from_pretrained("../visualglm-6b", trust_remote_code=True) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/transformers/models/auto/tokenization_auto.py", line 679, in from_pretrained return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1804, in from_pretrained return cls._from_pretrained( File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/transformers/tokenization_utils_base.py", line 1958, in _from_pretrained tokenizer = cls(*init_inputs, **init_kwargs) File "/home/lan/.cache/huggingface/modules/transformers_modules/visualglm-6b/tokenization_chatglm.py", line 221, in init self.sp_tokenizer = SPTokenizer(vocab_file, num_image_tokens=num_image_tokens) File "/home/lan/.cache/huggingface/modules/transformers_modules/visualglm-6b/tokenization_chatglm.py", line 64, in init self.text_tokenizer = TextTokenizer(vocab_file) File "/home/lan/.cache/huggingface/modules/transformers_modules/visualglm-6b/tokenization_chatglm.py", line 22, in init self.sp.Load(model_path) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/sentencepiece/init.py", line 905, in Load return self.LoadFromFile(model_file) File "/home/lan/anaconda3/envs/face19/lib/python3.9/site-packages/sentencepiece/init.py", line 310, in LoadFromFile return _sentencepiece.SentencePieceProcessor_LoadFromFile(self, arg) RuntimeError: Internal: src/sentencepiece_processor.cc(1101) [model_proto->ParseFromArray(serialized.data(), serialized.size())]

@wavelet2008 有解决吗 我也遇到同样的问题

参考一下这个:THUDM/ChatGLM-6B#770

是这样解决的,但chatGLM的清华仓库中是传了对应的ice_text.model文件,而visualGLM的清华仓库没上传最新的。。

qq281541534 avatar Jul 11 '23 08:07 qq281541534

有visualGLM模型的清华云盘的ice_text.model么?

Jerryyuanyuan avatar Jul 13 '23 08:07 Jerryyuanyuan