tempo-pytorch
tempo-pytorch copied to clipboard
Tempo-Pytorch on GPU Clusters
Hello @liaoyuhua Thanks for making the tempo-pytorch implementation available But I am facing a few issues while running it on a GPU cluster (Out of Memory issue), which says CUDA out of memory. Tried to allocate 20.00 MiB (GPU 0; 15.78 GiB total capacity; 13.53 GiB already allocated; 7.75 MiB free; 13.55 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
Can you please provide some information on how were you able to run this (compute details if possible) and the time taken for you to carry out the training process?
Thanks in advance for your help