layer-to-layer-pytorch
layer-to-layer-pytorch copied to clipboard
GPT-2 training example
Checklist
- [X] I've searched the project's
issues
.
❓ Question
Is it possible to provide an example of training GPT-2 with L2L?
📎 Additional context
I'm currently using the HuggingFace training script for train the GPT-2 model here.
Hello @loretoparisi, thank you for your interest in our work!
If this is a bug report, please provide screenshots and minimum viable code to reproduce your issue, otherwise we can not help you.
Hi! In HuggingFace models, layers are not in one ModuleList as required by L2L (and, for example, DeepSpeed).
But I will try to make such an example with GPT-2 👍
Hi! In HuggingFace models, layers are not in one ModuleList as required by L2L (and, for example, DeepSpeed).
But I will try to make such an example with GPT-2 👍
Will be good to make a translator for HuggingFace models into the L2L ModuleList format.