GRU4Rec_TensorFlow icon indicating copy to clipboard operation
GRU4Rec_TensorFlow copied to clipboard

A question about the testing process.

Open ghost opened this issue 7 years ago • 2 comments

Hello, Song,

I found a problem in this kind of process. In your evaluation.py: image I note that you picked ">" here, but if we use tanh as the final activation function, many scores of this preds may be equal to 1 for each user. And that means many scores of items are in one position at the same time. But this trick is unreasonable, because we cannot rank these items. If you use >=, the performance will be bad.

I think this is the implement in the original Hidash's code. So what do you think of this?

Looking forward to your reply. Thank you ~

ghost avatar Feb 11 '18 03:02 ghost

Thanks for discussing this! You are right this is the implement in the original Hidash's code. While I agree that using tanh as final activation function maybe problematic, especially together with cross entropy loss (negative number in log()). I only run this code with softmax activation and cross entropy loss. Would you please share a bit more on how you use tanh in your experiments? With what kind of loss function? How the performance is going?

Songweiping avatar Mar 06 '18 03:03 Songweiping

The original paper said that it is better to use tanh. 1557557837(1)

iejiaochenhao avatar May 11 '19 07:05 iejiaochenhao