litgpt icon indicating copy to clipboard operation
litgpt copied to clipboard

micro_batch_size, step run time, total training time

Open AngainorDev opened this issue 2 years ago • 0 comments

Hi,

Thanks a lot for this clear and fat-free code base! I'm training Falcon-7B with adapters-v2 and an Alpaca-formated dataset of mine.

As usual, I'm trying to max out the vram use for best training time but in this case, there is no significant gain since the step time is almost proportional to the batch size.

step times: micro_batch_size 1, 159ms micro_batch_size 2, 293ms micro_batch_size 4, 560ms

Is this expected, or can this be optimized?

Note: I'll also open a new issue as advised with my attempt at batch inference, exhibiting the same lack of gain when batching at inference, see https://github.com/Lightning-AI/lit-llama/issues/188#issuecomment-1586096141

AngainorDev avatar Jun 12 '23 06:06 AngainorDev