RICKand-MORTY
RICKand-MORTY
pytorch和CUDA版本不对应吧?
suitable flash-attn version: **1.0.2** or **0.2.8** I finally solve the problem.I changed my flash-attn version to **1.0.2** and **0.2.8** to test and both of these version is suitable.Besides, when I...
In cm/logger.py/configure you can find the following code: ` if dir is None: dir = os.getenv("OPENAI_LOGDIR") ` the output data was specified by environment variable “OPENAI_LOGDIR”.If not given, the output...
Maybe the version of pytorch or cuda is incorrect
是多卡训练吗?多卡训练dist_utils.py那个节点gpu数要改成自己的gpu数,另外命令行的mpiexec -n 4的4也要换成自己的gpu数
添加环境变量RDMAV_FORK_SAFE吧看看,可能是为了安全不让直接fork子进程 https://docs.nvidia.com/networking/display/rdmaawareprogrammingv17/ibv_fork_init
> > 添加环境变量RDMAV_FORK_SAFE吧看看,可能是为了安全不让直接fork子进程 https://docs.nvidia.com/networking/display/rdmaawareprogrammingv17/ibv_fork_init > > 在cm.train文件里添加了,但还是不行,报同样的错误 在/etc/profile里添加,作为系统环境变量
> 在/etc/profile里添加,作为系统环境变量 记得保存后用source刷新一下
is a package.use `pip install flash_attn==0.2.8` to install
Maybe your environment is not suitable. Here is my environment: python:3.11.5 torch:2.0.0 torchvision: 0.15.1 flash-attn:0.2.8 If still has problem in installing flash_attn,try installing manually,see here:https://github.com/Dao-AILab/flash-attention