opencompass
opencompass copied to clipboard
[Bug] TypeError: Qwen2ForCausalLM.__init__() got an unexpected keyword argument 'gpu_memory_utilization'
Prerequisite
- [X] I have searched Issues and Discussions but cannot get the expected help.
- [X] The bug has not been fixed in the latest version.
Type
I'm evaluating with the officially supported tasks/models/datasets.
Environment
{'CUDA available': True,
'CUDA_HOME': '/usr/local/cuda',
'GCC': 'gcc (Debian 12.2.0-14) 12.2.0',
'GPU 0': 'Tesla V100-PCIE-32GB',
'MMEngine': '0.10.4',
'MUSA available': False,
'NVCC': 'Cuda compilation tools, release 12.4, V12.4.99',
'OpenCV': '4.10.0',
'PyTorch': '2.3.0+cu121',
'PyTorch compiling details': 'PyTorch built with:\n'
' - GCC 9.3\n'
' - C++ Version: 201703\n'
' - Intel(R) oneAPI Math Kernel Library Version '
'2023.1-Product Build 20230303 for Intel(R) 64 '
'architecture applications\n'
' - Intel(R) MKL-DNN v3.3.6 (Git Hash '
'86e6af5974177e513fd3fee58425e1063e7f1361)\n'
' - OpenMP 201511 (a.k.a. OpenMP 4.5)\n'
' - LAPACK is enabled (usually provided by '
'MKL)\n'
' - NNPACK is enabled\n'
' - CPU capability usage: AVX512\n'
' - CUDA Runtime 12.1\n'
' - NVCC architecture flags: '
'-gencode;arch=compute_50,code=sm_50;-gencode;arch=compute_60,code=sm_60;-gencode;arch=compute_70,code=sm_70;-gencode;arch=compute_75,code=sm_75;-gencode;arch=compute_80,code=sm_80;-gencode;arch=compute_86,code=sm_86;-gencode;arch=compute_90,code=sm_90\n'
' - CuDNN 8.9.2\n'
' - Magma 2.6.1\n'
' - Build settings: BLAS_INFO=mkl, '
'BUILD_TYPE=Release, CUDA_VERSION=12.1, '
'CUDNN_VERSION=8.9.2, '
'CXX_COMPILER=/opt/rh/devtoolset-9/root/usr/bin/c++, '
'CXX_FLAGS= -D_GLIBCXX_USE_CXX11_ABI=0 '
'-fabi-version=11 -fvisibility-inlines-hidden '
'-DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO '
'-DLIBKINETO_NOROCTRACER -DUSE_FBGEMM '
'-DUSE_QNNPACK -DUSE_PYTORCH_QNNPACK '
'-DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE '
'-O2 -fPIC -Wall -Wextra -Werror=return-type '
'-Werror=non-virtual-dtor -Werror=bool-operation '
'-Wnarrowing -Wno-missing-field-initializers '
'-Wno-type-limits -Wno-array-bounds '
'-Wno-unknown-pragmas -Wno-unused-parameter '
'-Wno-unused-function -Wno-unused-result '
'-Wno-strict-overflow -Wno-strict-aliasing '
'-Wno-stringop-overflow -Wsuggest-override '
'-Wno-psabi -Wno-error=pedantic '
'-Wno-error=old-style-cast -Wno-missing-braces '
'-fdiagnostics-color=always -faligned-new '
'-Wno-unused-but-set-variable '
'-Wno-maybe-uninitialized -fno-math-errno '
'-fno-trapping-math -Werror=format '
'-Wno-stringop-overflow, LAPACK_INFO=mkl, '
'PERF_WITH_AVX=1, PERF_WITH_AVX2=1, '
'PERF_WITH_AVX512=1, TORCH_VERSION=2.3.0, '
'USE_CUDA=ON, USE_CUDNN=ON, USE_CUSPARSELT=1, '
'USE_EXCEPTION_PTR=1, USE_GFLAGS=OFF, '
'USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=ON, '
'USE_MKLDNN=ON, USE_MPI=OFF, USE_NCCL=1, '
'USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, '
'USE_ROCM_KERNEL_ASSERT=OFF, \n',
'Python': '3.10.14 (main, May 6 2024, 19:42:50) [GCC 11.2.0]',
'TorchVision': '0.18.0+cu121',
'numpy_random_seed': 2147483648,
'opencompass': '0.2.6+a62c613',
'sys.platform': 'linux'}
Reproduces the problem - code/configuration sample
MKL_SERVICE_FORCE_INTEL=0 \
MKL_THREADING_LAYER=GNU \
python run.py \
--hf-type chat \
--hf-path /data/public/qim/model/Qwen1.5-4B-Chat \
--model-kwargs gpu_memory_utilization=0.7 \
--tokenizer-path /data/public/qim/model/Qwen1.5-4B-Chat \
--generation-kwargs do_sample=True \
--accelerator vllm \
--work-dir /data/public/qim/model/opencompass/outputs \
--max-num-workers 10 \
--datasets Xsum_gen \
--debug
Reproduces the problem - command or script
MKL_SERVICE_FORCE_INTEL=0 \
MKL_THREADING_LAYER=GNU \
python run.py \
--hf-type chat \
--hf-path /data/public/qim/model/Qwen1.5-4B-Chat \
--model-kwargs gpu_memory_utilization=0.7 \
--tokenizer-path /data/public/qim/model/Qwen1.5-4B-Chat \
--generation-kwargs do_sample=True \
--accelerator vllm \
--work-dir /data/public/qim/model/opencompass/outputs \
--max-num-workers 10 \
--datasets Xsum_gen \
--debug
Reproduces the problem - error message
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
Traceback (most recent call last):
File "/data/public/qim/script/opencompass/opencompass/tasks/openicl_infer.py", line 161, in <module>
inferencer.run()
File "/data/public/qim/script/opencompass/opencompass/tasks/openicl_infer.py", line 73, in run
self.model = build_model_from_cfg(model_cfg)
File "/root/miniconda3/envs/swift-env/lib/python3.10/site-packages/opencompass/utils/build.py", line 25, in build_model_from_cfg
return MODELS.build(model_cfg)
File "/root/miniconda3/envs/swift-env/lib/python3.10/site-packages/mmengine/registry/registry.py", line 570, in build
return self.build_func(cfg, *args, **kwargs, registry=self)
File "/root/miniconda3/envs/swift-env/lib/python3.10/site-packages/mmengine/registry/build_functions.py", line 121, in build_from_cfg
obj = obj_cls(**args) # type: ignore
File "/root/miniconda3/envs/swift-env/lib/python3.10/site-packages/opencompass/models/huggingface_above_v4_33.py", line 169, in __init__
self._load_model(path=path, kwargs=model_kwargs, peft_path=peft_path, peft_kwargs=peft_kwargs)
File "/root/miniconda3/envs/swift-env/lib/python3.10/site-packages/opencompass/models/huggingface_above_v4_33.py", line 218, in _load_model
self.model = AutoModelForCausalLM.from_pretrained(path, **model_kwargs)
File "/root/miniconda3/envs/swift-env/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained
return model_class.from_pretrained(
File "/root/miniconda3/envs/swift-env/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3710, in from_pretrained
model = cls(config, *model_args, **model_kwargs)
TypeError: Qwen2ForCausalLM.__init__() got an unexpected keyword argument 'gpu_memory_utilization'
Other information
No response