COMET
COMET copied to clipboard
Batch size in words
🚀 Feature
being able to specify the batch size in terms of tokens could allow larger batch sizes.
Motivation
When training on data with various sentence lengths, increasing a batch size can break the training (especially when there is a language that is tokenized to too many tokens). Specifying batch size in terms of tokens, could fix this issue and allow better usage of the GPU