deep-coref icon indicating copy to clipboard operation
deep-coref copied to clipboard

Memory requirement for training on the conll-2012 corpus

Open thomwolf opened this issue 7 years ago • 2 comments

Hi, I am trying to train your model on a AWS p2x instance (with a 12 Go K80 GPU) on the Conll-2012 corpus (2802 documents in the training dataset). The training eats all (RAM) memory (64 Go) in less than 30 % of the first epoch and gets killed before finishing it.

I was wondering on what type of machine you trained it ? Is 64 Go of RAM too small for training on the conll corpus ?

thomwolf avatar May 17 '17 21:05 thomwolf

+1

julien-c avatar Jun 12 '17 09:06 julien-c

Huh, that strange. I trained the model on a 128G machine, but I don't think it should use more than 10G RAM. Do you know what's taking up all the memory? What is the size of the data/features directory created during preprocessing?

clarkkev avatar Jun 22 '17 05:06 clarkkev