Skip to content

[KDD 2022] Comprehensive Fair Meta-learned Recommender System

Notifications You must be signed in to change notification settings

weitianxin/CLOVER

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

17 Commits
 
 
 
 
 
 
 
 

Repository files navigation

CLOVER

This is an implemention for our SIGKDD 2022 paper based on Pytorch.

Comprehensive Fair Meta-learned Recommender System

by Tianxin Wei, Jingrui He.

Introduction

It's a general framework to comprehensively mitigate the unfairness issues in the cold-start recommender systems.

Requirements

  • python==3.6
  • torch==1.9.0
  • numpy==1.19.5
  • tqdm==4.64.0
  • scikit-learn==0.24.2

Other versions of these packages may also work.

Instructions

The code and data for the three datasets are placed in the three folders: ML1M, BX and ML100K.

For ML1M, we use the same dataset as in MELU. For the other two datasets, as there are no available datasets, we preprocess them ourselves following the same protocol. The processing python script of the original data is placed in the data folder.

Next we show how to run our code, first please enter the folder. Create the folder "log" to save the training log. Create the folder "multi_result_files" to save the trained model.

For the basic meta-learned recommender system, run:

python maml.py --cuda 0 --rerun --num_epoch 50 --loss 1 --seed 53

For our fair adversarial meta-learned recommende system, run:

# just the first adversarial loss function without external information
python maml.py --cuda 0 --rerun --adv 1 --fm --num_epoch 50 --loss 1 --out 0 --outer 0 --inner_fc 0 --seed 53 --disable_inner_max

# just the first adversarial loss function with external information
python maml.py --cuda 0 --rerun --adv 2 --fm --num_epoch 50 --loss 1 --out 5 --outer 0 --inner_fc 0 --seed 53 --disable_inner_max

# the two adversarial loss function together
python maml.py --cuda 0 --rerun --adv 2 --adv2 1 --fm --num_epoch 50 --loss 1 --out 5 --out2 2 --outer 0 --inner_fc 0 --seed 53 --disable_inner_max --normalize --item_adv

The hyperparameters need to be further finetuned for each dataset to obtain better performance. In these commands, "adv" denotes the tradeoff factor of the first adversarial loss and adv2 represents the second. "loss=1" denotes use the cross-entropy loss. The argument "out" controls the input information to the adversarial network. "disable_inner_max" is to disable the optimization of task T_2 in the inner loop. "disable_inner_adv" is to disable the optimization of task T_1.

The final results are the average of five independent runs (seeds) with the best model. For the AUC metric, it is tested using a trained model to infer the sensitive attribute with function "fair_fine" in the "maml.py" file.

Acknowledges

Part of our code references the MeLU_pytorch code repo. Many thanks for their public code.

Citation

If you find this paper or code helpful, please consider cite our paper. Thanks very much.

@inproceedings{wei2022comprehensive,
  title={Comprehensive Fair Meta-learned Recommender System},
  author={Wei, Tianxin and He, Jingrui},
  booktitle={Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining},
  pages={1989--1999},
  year={2022}
}

About

[KDD 2022] Comprehensive Fair Meta-learned Recommender System

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Languages