Woo-Seong YunSeongmin KangYoon-Sik Cho
Recommender systems aim to accurately predict user preferences in order to provide potential items of interests. However, the highly skewed long-tail item distribution leads the models to more focus on popular items, which can damage the predictive performance by recommending specific items repeatedly. We propose a novel Variational Autoencoder (VAE) framework for collaborative filtering using contrastive disentanglement. We contrast salient latent features in VAE against the non-salient background. Here, we intentionally generate a background dataset based on item popularity, a concept previously unexplored in existing recommender system. The representations learnt from our proposed scheme better reflect salient latent factor instead of being washed out by latent factor of popular items. Consequently, our model improves the predictive performance by effectively learning the representations toward salient latent features while excluding the effects from popularity. The proposed contrastive disentanglement framework is generic and thus adaptable to VAE-based recommender algorithms. We comprehensively evaluate our proposed method on popular datasets such as MovieLens20M and Netflix and show that it consistently outperforms corresponding VAE models, achieving superior performance across all metrics. Notably, in terms of diversity, our approach outperforms the best-performing methods by 5.5% and 4.8% on the MovieLens20M and Netflix datasets, respectively. We further present a way of extending the framework to an ensemble of two contrasts with two different backgrounds, which achieves state-of-the-art performance in standard settings with significant improvements. Our code is available at https://github.com/yunwooseong/CD-VAE
WU Mei-lin, HUANG Jia-jin, QIN Jin
Xubin RenLianghao XiaJiashu ZhaoDawei YinChao Huang
Sujie YuJunnan ZhuoLvying ChenHailian YinBohan Li
Y. YuanZhen LiuXinxin YangSibo Lu
Quoc-Tuan TruongAghiles SalahHady W. Lauw