Kathan Shah

Results 3 issues of Kathan Shah

1. Added epsilon in denominator to the normalization in the encoder, because it was throwing nans otherwise. 2. Added check in NDCG calculation to skip users for which there were...

Hi, Just another suggestion. Wouldn't the more accurate name for the classes Attention and LinearAttention in rin_pytorch.py be CrossAttention and SelfAttention?

Hi, I was looking to implement the RIN architecture using your code. I noticed you applied positional embedding in every block, as opposed to just once in the beginning. What...