Chinmayrane16 / DeepRecommender

Training Deep AutoEncoders for Collaborative Filtering

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Deep AutoEncoders for Collaborative Filtering Tweet

License: MIT

This repository is an implementation of the research paper by NVIDIA :

Training Deep AutoEncoders for Collaborative Filtering
- Oleksii Kuchaiev and Boris Ginsburg
[arXiv][Github]

Collaborative Filtering

Collaborative Filtering is a method used by recommender systems to make predictions about an interest of an specific user by collecting taste or preferences information from many other users. The technique of Collaborative Filtering has the underlying assumption that if a user A has the same taste or opinion on an issue as the person B, A is more likely to have B’s opinion on a different issue.

Deep AutoEncoder

Autoencoder: AutoEncoders are a type of Feed forward neural networks that consits of 3 different layers - encoding layer, representation layer and decoding layer. The representation layer is also called the coding layer. An undercomplete autoencoder has less number of neurons in the coding layer in order to limit the flow of information through the network (bottleneck). An overcomplete autoencoder has greater number of neurons in the hidden layer than the input layer. So, representation layer consists of the latent representations of the input.

They are a form of unsupervised learning, as the input itself serves as the output, for the purpose of reconstructing it's inputs.
The following image is taken from the research paper.

Implementation

In this repository, I have implemented the state-of-the-art technique for predicting the user ratings using Deep AutoEncoders with the help of Pytorch framework.
I have used the Movielens 100k dataset.
You can find the dataset on this repository - ml-latest-small

Requirements

Dataset Preprocessing

The ratings.csv file consists of users and their ratings.

I have sorted the dataset based on Timestamp as mentioned in the paper. For the rating prediction task, it is often most relevant to predict future ratings given the past ones instead of predicting ratings missing at random. Training interval contains ratings which came in earlier than the ones from testing interval. Users and items that do not appear in the training set are removed from the validation sets.

This is done in the DataLoader.ipynb file.
You can find the sorted train and test datasets here - (train) (test)

Model Structure

The model has 7 layers having the layer sizes [n, 512, 512, 1024, 512, 512, n]. Some of the tuned hyperparamaters are as follows -

Hyperparameters Value Explanation
activation_function selu type of non-linearity
dropout 0.0 dropout probability
num of hidden neurons 1024 Number of neurons in each layer
learning_rate 0.001 learning rate

The model uses Adam as optimizer with lr = 0.001 although the paper suggests using SGD with lr = 0.001 and momentum = 0.9.
The loss function used is MMSE (Masked Mean Square Error) and to evaluate performance, I have used RMSE which is experimentally found to be square root of MMSE.

At the end of the Training for 40 epochs, I acheived RMSE training loss of 0.5567 and validation loss of 0.39.

HyperParameter Tuning

  • Activation Function

Trying different activation functions listed in the paper.

We can see that selu and elu functions perform much better than any other activation function, exactly as stated in the paper.

  • Dropout Probability

Tried 3 dropout probabilities - dp 0.0 , dp 0.25, dp 0.65

We can see that here since the data size is small therefore it's better to not use dropout.

  • Hidden Layer

Tried experimenting with different sizes of the representation layer. I found the hypothesis true, that more the number of neurons in the hidden layer more better the representation. But there happens to be a case where the layer simply tries to copy the data from previous layers if you have overcomplete autoencoders.

Below are the results i found -

  • Learning Rate

Tried different learning rates -

We cannot choose a learning rate too high otherwise the loss will go on fluctuating too rapidly and we cannot choose a learning rate too slow otherwise it will take lot of time to converge. Personally, I found learning rate 0.001 to be the best among all.

Conclusion

In this repository, I have used Movielens 100k ratings dataset. I have segregated the dataset into train and validation sets according to the timestamp. I have chosen a timestamp such that validation data consists of the ratings 2 months later which we need to predict.

I have used 7 layer autoencoder architecture (including the input and the output layer) as stated in the paper with 1024 hidden layer neurons. Moreover, I have found 'selu' activation function to perform better on this dataset. I have used Adam optimizer to adapt weights with a learning rate of 0.001 . I have tried using different dropout probabilities but found that the higher the dropout the worse the performance as opposed to higher dropout probability stated in the paper, I guess this contradicition comes from different sizes of our datasets. Finally, at the end of 40 epochs, I have acheived RMSE training loss of 0.5567 and validation loss of 0.39 .

References

https://arxiv.org/pdf/1708.01715v3.pdf
https://grouplens.org/datasets/movielens/
https://towardsdatascience.com/deep-autoencoders-for-collaborative-filtering-6cf8d25bbf1d
https://github.com/NVIDIA/DeepRecommender
https://github.com/shivamsaboo17
https://www.jeremyjordan.me/autoencoders/
https://www.youtube.com/platform.ai

About

Training Deep AutoEncoders for Collaborative Filtering

License:MIT License


Languages

Language:Jupyter Notebook 96.4%Language:Python 3.6%