CogView2
CogView2 copied to clipboard
Replicate CUDA out of memory
replicate fails after some images
CUDA out of memory. Tried to allocate 1.22 GiB (GPU 0; 39.59 GiB total capacity; 30.86 GiB already allocated; 1.07 GiB free; 34.03 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
I am getting same :{
Also the same for me. About 1/5 runs actually generates images. Very nice demo though. Thank you for making the work available. Good luck!
Hi, I have checked the CUDA usage locally, it does seem some memory persists after a predict()
run, I have pushed to the website a new version freeing about 10G cached memory between consecutive inferences, which should solve the issue. Have a try whether it helps!
https://replicate.com/p/ih4rc5rid5cu3km25xzki4rk3q
How much VRAM is needed for this to run?
Tried to allocate 72.00 MiB (GPU 0; 23.68 GiB total capacity; 21.73 GiB already allocated; 46.31 MiB free; 21.74 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
I tried to reduce the --max-inference-batch-size 8
Honestly, this AI Model is entirely useless if the Demo System can't even produce a SINGLE Image... And this is based on around 400 attempts...