MiniCPM-V icon indicating copy to clipboard operation
MiniCPM-V copied to clipboard

MiniCPM-Llama3-V 2.5 全参数微调最小的显存需求

Open Millie-Xu opened this issue 1 year ago • 4 comments

是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?

  • [X] 我已经搜索过已有的issues和讨论 | I have searched the existing issues / discussions

该问题是否在FAQ中有解答? | Is there an existing answer for this in FAQ?

  • [X] 我已经搜索过FAQ | I have searched FAQ

当前行为 | Current Behavior

No response

期望行为 | Expected Behavior

针对 MiniCPM-Llama3-V 2.5,我使用两张2张v100或者4张v100,batchsize设置为1,deepspeed zero3配置,并尽可能将所有参数放置在cpu上,仍然显示显存不够,请问该模型全参数微调所需的最小显存需求是什么?

复现方法 | Steps To Reproduce

No response

运行环境 | Environment

- OS:
- Python:
- Transformers:
- PyTorch:
- CUDA (`python -c 'import torch; print(torch.version.cuda)'`):

备注 | Anything else?

No response

Millie-Xu avatar Jun 20 '24 02:06 Millie-Xu

@Millie-Xu offload到cpu都不够吗?另外请问您使用的是什么微调脚本呢?

yihp avatar Jun 25 '24 12:06 yihp

你好,我已经解决啦,谢谢

Millie-Xu avatar Jun 27 '24 03:06 Millie-Xu

你好,我已经解决啦,谢谢 请问你是怎么解决的呢?最终占用了多少显存?

todaydeath avatar Jun 29 '24 05:06 todaydeath

你好,我已经解决啦,谢谢 请问你是怎么解决的呢?最终占用了多少显存?

用了8张V100,显存差不多占满了

Millie-Xu avatar Jul 01 '24 02:07 Millie-Xu

那你应该是全量微调对吧

LDLINGLINGLING avatar Jul 04 '24 08:07 LDLINGLINGLING