llama-recipes
llama-recipes copied to clipboard
Add support to continue pratraining
🚀 The feature, motivation and pitch
Currently the LLaMA-receipts do not support pretraining out of box. For example, one could not save the pretrained model based on the current checkpoint saving logic which requires run_validation = True, which is not needed during continued pretraining phase. It would be nice to add this feature down the road as it's a common use case in the process of domain adaption.
Thanks.
Alternatives
No response
Additional context
No response
@hanyin88 Did you find anything for this elsewhere?
@hanyin88 the original focus of this repo was about fine-tuning, so it has not been designed to support pertaining, but this is something we are investigating as well.