Luca Antiga

Results 171 comments of Luca Antiga

Great! @Diormiu we'll get this merged as soon as the fix gets, in. If you don't have time we can push this through no problem.

Hey @mfranzon, that would be cool! Is it something you'd be interested in contributing?

Changing batch size will not change the memory requirements, since we are using gradient accumulation, but changing `micro_batch_size` will. What happens is that forward / backward will be computed with...

Sorry what do you mean by "multiple-lan support"?

Hi, can you post the CLI args or code you are using? Also this is with two machines and 8 GPUs per machine?

Just to confirm: are you running the pretraining command? Maybe try to comment this line out: https://github.com/Lightning-AI/litgpt/blob/main/litgpt/pretrain.py#L174 We have bumped into issues with PyTorch 2.2 and torch.compile recently, let's take...

Thanks for the report. Can you try: - running without torch.compile (comment this line out https://github.com/Lightning-AI/litgpt/blob/main/litgpt/pretrain.py#L174) - running with torch.compile but on PyTorch 2.3 Thanks a lot for investigating this....

Hey @khushi-411 let us know if you need help!