xiangriconglin
xiangriconglin
> For CNN/DM dataset with 287226 train size, I'd suggest the following setup: > Train this model using batch size 32 for 15 epochs (max_iter=134637) > then activate RL training...
> **my command is blow:** > `python run_summarization.py --mode=decode --data_path=../data_no_extract/finished_files/chunked/submit_* --vocab_path=../data_no_extract/finished_files/vocab --log_root=../log --exp_name=intradecoder-temporalattention-withpretraining --rl_training=False --intradecoder=True --use_temporal_attention=True --single_pass=1 --beam_size=4 --decode_from=eval` > I just want use pretraining model to decode, so the...
Thank you very much. The issue has been solve. Thank you for your reply.
> No, this issue is [well-discussed](https://github.com/abisee/pointer-generator/issues/12) in the original pointer-generator model page. > Every time you run this model, it will generate a different result due to the multi-processing batching...