MiniCPM icon indicating copy to clipboard operation
MiniCPM copied to clipboard

[Bad Case]: 使用minicpm4-0.5B和8B的模型都报错

Open Yhc-777 opened this issue 2 months ago • 1 comments

Description / 描述

你好,请问我在将minicpm4-0.5B和minicpm4-8B的模型下载下来之后,使用范例代码推理: from transformers import AutoModelForCausalLM, AutoTokenizer import torch torch.manual_seed(0)

path = "/mnt/2/haochen/LLM/MiniCPM/pretrained_models/MiniCPM4-0.5B" device = "cuda" tokenizer = AutoTokenizer.from_pretrained(path) model = AutoModelForCausalLM.from_pretrained(path, torch_dtype=torch.bfloat16, device_map=device, trust_remote_code=True)

responds, history = model.chat(tokenizer, "Write an article about Artificial Intelligence.", temperature=0.7, top_p=0.7) print(responds)

均报错如下: (llm_minicpm4) u@u:/mnt/2/haochen/LLM/MiniCPM$ python transformers_inference_small.py Traceback (most recent call last): File "/mnt/2/haochen/LLM/MiniCPM/transformers_inference_small.py", line 7, in <module> tokenizer = AutoTokenizer.from_pretrained(path) File "/mnt/2/haochen/envs/llm_minicpm4/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py", line 880, in from_pretrained return tokenizer_class.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs) File "/mnt/2/haochen/envs/llm_minicpm4/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 2110, in from_pretrained return cls._from_pretrained( File "/mnt/2/haochen/envs/llm_minicpm4/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 2336, in _from_pretrained tokenizer = cls(*init_inputs, **init_kwargs) File "/mnt/2/haochen/envs/llm_minicpm4/lib/python3.10/site-packages/transformers/models/llama/tokenization_llama_fast.py", line 159, in __init__ super().__init__( File "/mnt/2/haochen/envs/llm_minicpm4/lib/python3.10/site-packages/transformers/tokenization_utils_fast.py", line 114, in __init__ fast_tokenizer = TokenizerFast.from_file(fast_tokenizer_file) Exception: data did not match any variant of untagged enum ModelWrapper at line 490842 column 3

但是我用MiniCPM-2B-sft-bf16就可以正常inference,请问该如何解决这个问题呢?谢谢!

Case Explaination / 案例解释

No response

Yhc-777 avatar Sep 25 '25 02:09 Yhc-777

这可能是由于tokenizers和transformers库的版本不一致导致的。你可以尝试更新tokenizers库。如果您使用的是MiniCPM4进行推理,可以尝试将tokenizers更新至版本0.20.0。如果问题仍然存在,请把相关库的信息提供给我。

hansjohn avatar Sep 26 '25 07:09 hansjohn