variational-autoencoders-for-collaborative-filtering-pytorch icon indicating copy to clipboard operation
variational-autoencoders-for-collaborative-filtering-pytorch copied to clipboard

Implementation of variational autoencoders for collaborative filtering in PyTorch

Variational autoencoders for collaborative filtering in PyTorch.

This repo implements Variational autoencoders for collaborative filtering in PyTorch presented in [1], and also does conditional VAE [2] to use user profiles in collaborative filtering. Numerical expriments are done for MovieLens dataset for VAE and Last.fm for conditional VAE.

Dataset

Two datasets are used in this repo.

  • MovieLens 20M Dataset: This contains rating for movies by users. This collaborative filtering recommends movies user would like to watch.
  • Last.fm Dataset - 360K users: This contains play counts of artists by users and user profile, gender, age, country and signup data. This collaborative filtering recommends artists user would like to listen to, using play counts and user profiles. You can use VAE conditioned on the the user profiles.

After downloading datasets, run

create_data.sh

Training

Usage:

python demo.py train <options>

Results

model data NDCG@50(validation) NDCG@50(test)
DAE(TensorFlow) MovieLens 20M 0.42778 0.42113
DAE(PyTorch) MovieLens 20M 0.42574 0.41932
VAE(TensorFlow) MovieLens 20M 0.43340 0.42593
VAE(PyTorch) MovieLens 20M 0.43887 0.43093
VAE(PyTorch) Last.fm 360K 0.39175 0.39152
VAE(PyTorch) conditioned on gender, age and country Last.fm 360K 0.38344 0.38346

NDCG@N and Recall@N of VAE(PyTorch) for MovieLens 20M

NDCG@N and Recall@N

References

  1. Dawen Liang, Rahul G. Krishnan, Matthew D. Hoffman, Tony Jebara. Variational Autoencoders for Collaborative Filtering, The Web Conference (WWW), 2018.
    arXiv, github

  2. Kihyuk Sohn, Honglak Lee, and Xinchen Yan. Learning Structured Output Representation Using Deep Conditional Generative Models. Advances in Neural Information Processing Systems, 2015. NIPS

  3. Irina Higgins, Loic Matthey, Arka Pal, Christopher Burgess, Xavier Glorot, Matthew Botvinick, Shakir Mohamed, Alexander Lerchner. beta-VAE: Learning basic visual concepts with a constrained variational framework. ICLR, 2017. Paper link