opencompass
opencompass copied to clipboard
[Bug] I downloaded the Llama-7b model on the huggingface to the local location, and then modified the path to the file location. After running the code, the following error occurred. How can I solve it?
先决条件
问题类型
我正在使用官方支持的任务/模型/数据集进行评估。
环境
python
重现问题 - 代码/配置示例
python run.py --datasets ceval_ppl --hf-path /root/pruning/llama-7b --tokenizer-kwargs padding_side='left' truncation='left' trust_remote_code=True --model-kwargs device_map='auto' --max-seq-len 2048 --max-out-len 100 --batch-size 64 --num-gpus 1
重现问题 - 命令或脚本
python run.py --datasets ceval_ppl --hf-path /root/pruning/llama-7b --tokenizer-kwargs padding_side='left' truncation='left' trust_remote_code=True --model-kwargs device_map='auto' --max-seq-len 2048 --max-out-len 100 --batch-size 1--num-gpus 1
重现问题 - 错误信息
其他信息
No response
Please show us the content of outputs/blabla/logs/infer/blabla/blabla.out and outputs/blabla/logs/eval/blabla/blabla.out
GPU: A100 40GB
Loading checkpoint shards: 0%| | 0/2 [00:00<?, ?it/s]ERROR:torch.distributed.elastic.multiprocessing.api:failed (exitcode: -9) local_rank: 0 (pid: 626117) of binary: /root/.local/conda/envs/pytorch/bin/python
Traceback (most recent call last):
File "/root/.local/bin/torchrun", line 8, in
I have the same problem (OpenICLEval fail), can you solve it?
Please show us the content of
outputs/blabla/logs/infer/blabla/blabla.outandoutputs/blabla/logs/eval/blabla/blabla.out
'torchrun' 不是内部或外部命令,也不是可运行的程序或批处理文件。
同问,我的报错是/bin/sh: torchrun: command not found,查了下说torch1.9.1之后就支持torchrun了,我是torch2.0.1,还是报错