deep-learning-benchmark
deep-learning-benchmark copied to clipboard
Minibatch size when going to mixed precision
Thank you for excellent data points!
Can you estimate potential increase in minibatch size when going to mixed precision?
Nvidia claims memory usage should go down, but aren't specific.
In my experiments with Titan V (using Tensorflow and home-grown implementation of Transformer model) I can only increase batch size by about 10%, which is much less than I expected.
Thanks!