tensor2tensor icon indicating copy to clipboard operation
tensor2tensor copied to clipboard

pretrained_model_dir not implemented

Open ekuznetsov139 opened this issue 5 years ago • 5 comments

layers/common_hparams.py mentions a hyperparameter "pretrained_model_dir":

"Directory containing a checkpoint for a pretrained model. This will only be used if a new run is being started. Parameters not found in the pretrained model will be randomly initialized. Superfluous parameters in the pretrained model will be ignored."

It sounds like it could be useful, however, as far as I can tell, the hyperparameter is not actually hooked up anywhere. The name of the parameter does not appear anywhere else in the sources of tensor2tensor, tensorflow, or tensorflow_estimator.

Is this feature obsolete or not yet implemented?

ekuznetsov139 avatar Jul 24 '19 21:07 ekuznetsov139

maybe option warm_start_from is what you need;

dreamingo avatar Aug 05 '19 04:08 dreamingo

Possibly, but that's not the point.

ekuznetsov139 avatar Aug 05 '19 05:08 ekuznetsov139

@ekuznetsov139 @lukaszkaiser Any update on this?

sauravjoshi avatar Jun 01 '20 10:06 sauravjoshi

I don't believe this repo is active any more. From the readme, it is now in maintenance mode and they encourage people to use Trax instead. Though the last time I tried to use Trax, I had to give up because it lacked some essential features like the tokenizer, and it was not clear whether they expected me to use t2t for that or to write my own code, or they simply neglected to open source the necessary bits. Maybe things have changed since then.

ekuznetsov139 avatar Jun 01 '20 23:06 ekuznetsov139

@ekuznetsov139 Thanks pal, for the info and time.

sauravjoshi avatar Jun 02 '20 20:06 sauravjoshi