helsenIgnace

Results 2 comments of helsenIgnace

I had the same error, turned out an 'with torch.cuda.amp.autocast(dtype=torch.float32)'. Be sure to check whether you changed all the torch.cuda.amp.autocast() to use torch.float32 instead of bfloat16.

When do you think this pr could be merged?