litgpt
litgpt copied to clipboard
Chunked LM head for lower peak memory during finetuning
Proposed by @robieta
I removed the lora context manager in favor of a separate model to implement this, just as we do for adapter.