Kai LiLong XuCheng ZhuKunlun Zhang
Research on recommendation methods using multimodal graph information presents a significant challenge within the realm of information services. Prior studies in this area have lacked precision in the purification and denoising of multimodal information and have insufficiently explored fusion methods. We introduce a multimodal graph recommendation approach leveraging cross-attention fusion. This model enhances and purifies multimodal information by embedding the IDs of items and their corresponding interactive users, thereby optimizing the utilization of such information. To facilitate better integration, we propose a cross-attention mechanism-based multimodal information fusion method, which effectively processes and merges related and differential information across modalities. Experimental results on three public datasets indicated that our model performed exceptionally well, demonstrating its efficacy in leveraging multimodal information.
Peishan LiWeixiao ZhanLutao GaoShuran WangLinnan Yang
Weijian HuangJianhua WuWeihu SongZehua Wang
ZHANG Xiaohui, MA Huifang, WANG Wentao, GAO Zihao