modded-nanogpt
modded-nanogpt copied to clipboard
Inference
how to do inference?
Did you post the final models?
I did a straight run of this repo, unmodified (except for the obvious run parameters), on the cloud using one L40S. Results with the checkpoints/log/final model/inference code/online demo is posted at huggingface here: https://huggingface.co/lemonteaa/nanogpt-speedrun
(No claim that the codes are bug free though as I just ask AI to code it up quickly with some manual coding on the non core part)
I think gpt-fast will do a lot of help