ray0809
ray0809
@dipanjan06 In [https://github.com/CharlesShang/FastMaskRCNN/blob/master/train/train.py#L230](url) May be a lower fraction can solve your problem
@sairin1202 Would you please share the 'checkpoint' file after 500K iterations?
@Fazziekey thanks for your response In the colossalai's stable diffusion example, the flash_attn's tag is disable, but the readme tells us that using flash could save much gpu memory. So...
ok, Thank you~
@Fazziekey Could you tell me, which branch are you developing on? my env: Cuda version: 11.3 Pytorch: 1.12.0 pytorch-lightning: 1.9.0.dev0 (from 1SAA git) colossalai: 0.1.10+torch1.12cu11.3 (compile from source, 0.1.11, 0.1.12,...
[2114#issuecomment-1361721810](https://github.com/hpcaitech/ColossalAI/issues/2114#issuecomment-1361721810) I found out that the problem was caused because of xformer, when I turned it off, the finetuning code could be run
If `cond_stage_trainable = True`, it will also report an error ```python │ /opt/conda/lib/python3.7/site-packages/colossalai/gemini/chunk/manager.py:159 in get_chunk │ │ 156 Args: │ 157 tensor (torch.Tensor): a torch tensor object │ 158 """...
@riturajsoni @chrisjyw cd text_extraction cmake CMakeLists.txt make then will produce a bin file named 'text_extraction' ./text_extraction your_pic_path