nanoGPT
nanoGPT copied to clipboard
dropout is 0.0
The dropout of the GPT model in GPTConfig class is set to 0.0, this means there won't be any dropout when training though, correct?
Yes, by deafult
Many recent researches on LLMs proved that it's ok to not do dropout in pertaining. But you normally want dropout in fine tuning to avoid overfitting.
Yes, Absolutely core because dropout of 0.0 during training shows that there is no dropout during training