Skip to content

Latest commit

 

History

History
126 lines (92 loc) · 8.22 KB

README.md

File metadata and controls

126 lines (92 loc) · 8.22 KB

Leveraging Content-Style Item Representation for Visual Recommendation

This is the official GitHub repository for the paper Leveraging Content-Style Item Representation for Visual Recommendation, accepted as short paper at the 44th European Conference on Information Retrieval (ECIR 2022).

Authors: Yashar Deldjoo, Tommaso Di Noia, Daniele Malitesta*, Felice Antonio Merra.
*corresponding author

If you want to use our model as baseline in your paper, please remember to cite us:

@inproceedings{DBLP:conf/ecir/DeldjooNMM22,
  author    = {Yashar Deldjoo and
               Tommaso {Di Noia} and
               Daniele Malitesta and
               Felice Antonio Merra},
  title     = {Leveraging Content-Style Item Representation for Visual Recommendation},
  booktitle = {{ECIR} {(2)}},
  series    = {Lecture Notes in Computer Science},
  volume    = {13186},
  pages     = {84--92},
  publisher = {Springer},
  year      = {2022}
}

Disclaimer #1

The codes for the proposed model and the baselines were implemented in Elliot. If you want to reproduce the exact same results described in the paper, please consider this essential version of the framework. However, please, refer to the most updated version of the official GitHub page and the documentation for detailed information on how to run Elliot.

Disclaimer #2

We are working to integrate the proposed model within the most recent version of Elliot, so that we can take advantage of all the new introduced features. When this step is complete, we will have this repository directly linking to Elliot.

Table of Contents

Requirements

To begin with, please make sure your system has these installed:

  • Python 3.6.8
  • CUDA 10.1
  • cuDNN 7.6.4

Then, install all required Python dependencies with the command:

pip install -r requirements.txt

Training and Evaluation

For each experiment, please run the following script:

python start_experiment.py --config <configuration_filename_without_extension>

This will train or test all the models on the considered datasets, following the configuration files in ./config_files/ (where you can find a detailed overview on all explored values for all the hyperparameters).

Datasets

Dataset # Users # Items # Interactions
Amazon Boys & Girls* 1,425 5,019 9,213
Amazon Men* 16,278 31,750 113,106

* https://jmcauley.ucsd.edu/data/amazon/

The two adopted datasets are sub-categories of Amazon's product category Clothing, Shoes and Jewelry. We used the 2014 version (refer to the official link above). As for the filtering phase, we considered only the interactions recorded after 2010, and filtered out items and users with less than 5 interactions (applying the 5-core techniques on items and users). Finally, we applied the temporal leave-one-out to split the dataset into train, validation, and test sets as described in the paper.

After downloading the zip files pointed by the table links, you will have access to the following data:

  • train, validation, and test sets
  • a file to map visual features to items (to be removed when the code is integrated within the new version of Elliot)
  • the mapping between users and items ids and their original Amazon unique codes
  • the extracted visual features to train and evaluate the baselines and the proposed model

If you do not want to modify the configuration files, then you need to create a folder ./data/ where you will place the two downloaded datasets. Otherwise, feel free to change the paths from the configuration files and use your own setting.

Disclaimer. As we do not own the dataset, we decided not to release the original product images (you would need them to train DVBPR, or to extract each of the visual features - we already extracted them for you, though). Once again, we encourage you refer to the official link to get to the full dataset, and use the item mapping we provided to download again the original product images, if needed.

Visual Features. All visual features were extracted using the scripts accessible at this repository. The repository is maintained by our research group, and its purpose is to provide a comprehensive framework to extract both handcrafted and trainable features from multimedia data, e.g., images and texts. Please, feel free to join us and contribute to it, since its developing is still open and highly active!

Baselines and Our Method

Traditional Collaborative Filtering

Model Paper
BPRMF Rendle et al.
NeuMF He et al.

Visual-Based Collaborative Filtering

Model Paper
VBPR He and McAuley
DeepStyle Liu et al.
DVBPR Kang et al.
ACF Chen et al.
VNPR Niu et al.

Our Model

You may find the scripts for our proposed method here. We renamed our model CSV, which stands for Content-Style Visual Recommendation.

Reproducibility Details

We randomly initialize the model parameters of all tested methods with a Gaussian distribution with a mean of 0 and standard deviation of 0.01 and set the latent factor dimension to 128 following the experimental settings used in Chen et al.. We explore the following hyperparameters via grid-search: the learning rate in {0.0001, 0.001, 0.01} and the regularization coefficients in {0.00001, 0.001}, whereas we fix the batch size to 256, and the temperature to 10.0. We adopt early-stopping to avoid overfitting and choose the best model configuration for each algorithm according to the hit ratio (HR), i.e., the validation metric, as in Chen et al..

That being said, you may refer to the two configuration files:

to have a detailed overview on all adopted hyperparameters, evaluation settings, and files to run the experiments. The files are in yml format (as required in Elliot), so their meaning should be easy to follow and understand.

Contacts

*corresponding author