Abstract
The realm of recommender systems has experienced a notable upswing in interest, particularly in the context of multi-modality, where user preferences are characterized through the integration of behavioral data and diverse modal information associated with items. However, existing methods grapple with two significant challenges: (1) The inherent noise present in multi-modal features can contaminate item representations. Conventional fusion methods may inadvertently propagate this noise to interaction data through the fusion process. (2) Existing multi-modal recommendation methods typically rely on random data augmentation. These approaches introduce noise manually and may not fully exploit the latent potential in multi-modal information. To bridge the gap, we propose a novel methodology for a comprehensive integration of both multi-modal features and collaborative signals, termed Modality-Guided Collaborative Filtering (MGCF). This method delves into self-supervised signals derived from both the structural and semantic information of the features. These signals are then utilized to select and mask critical interactions adaptively. In our pursuit of generating discriminative representations, we employ a masked auto-encoder to distill informative self-supervision signals. Simultaneously, we aggregate global information through the process of reconstructing the masked subgraph structures. We evaluate the effectiveness of MGCF through extensive experiments on real-world datasets and verify the superiority of our method for multi-modal recommendation over various state-of-the-art baselines.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Ge, Y., et al.: Understanding echo chambers in e-commerce recommender systems. In: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 2261–2270 (2020)
Lyu, Y., Yin, H., Liu, J., Liu, M.: Reliable recommendation with review-level explanations. In: 2021 IEEE 37th International Conference on Data Engineering (ICDE), pp. 1548–1558 (2021)
He, R., McAuley, J.: VBPR: visual Bayesian personalized ranking from implicit feedback. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 30 (2016)
Wei, Y., Wang, X., Nie, L., He, X., Hong, R., Chua, T.-S.: MMGCN: multi-modal graph convolution network for personalized recommendation of micro-video. In: Proceedings of the 27th ACM International Conference on Multimedia, pp. 1437–1445 (2019)
Tao, Z., et al.: Self-supervised learning for multimedia recommendation. IEEE Trans. Multimedia (2022)
Chen, X., He, K.: Exploring simple Siamese representation learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 15750–15758 (2021)
Tao, Z., Wei, Y., Wang, X., He, X., Huang, X., Chua, T.-S.: MGAT: multimodal graph attention network for recommendation. Inf. Process. Manag. 57, 102277 (2020)
Zhou, X., et al.: Bootstrap latent representations for multi-modal recommendation. In: Proceedings of the ACM Web Conference 2023, pp. 845–854 (2023)
He, K., Chen, X., Xie, S., Li, Y., Dollár, P., Girshick, R.: Masked autoencoders are scalable vision learners. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 16000–16009 (2022)
Sun, F.-Y., Hoffmann, J., Verma, V., Tang, J.: Infograph: unsupervised and semi-supervised graph-level representation learning via mutual information maximization. arXiv preprint arXiv:1908.01000 (2019)
Xia, L., Huang, C., Shi, J., Xu, Y.: Graph-less collaborative filtering. In: Proceedings of the ACM Web Conference 2023, pp. 17–27 (2023)
Tang, J., Qu, M., Wang, M., Zhang, M., Yan, J., Mei, Q.: Line: large-scale information network embedding. In: Proceedings of the 24th International Conference on World Wide Web, pp. 1067–1077 (2015)
He, X., Deng, K., Wang, X., Li, Y., Zhang, Y., Wang, M.: LightGCN: simplifying and powering graph convolution network for recommendation. In: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in INFORMATION Retrieval, pp. 639–648 (2020)
Velickovic, P., Cucurull, G., Casanova, A., Romero, A., Lio, P., Bengio, Y., Others: graph attention networks. stat 1050, 48510–48550 (2017)
Kitaev, N., Kaiser, Ł., Levskaya, A.: Reformer: the efficient transformer. arXiv preprint arXiv:2001.04451 (2020)
Wei, W., Huang, C., Xia, L., Zhang, C.: Multi-modal self-supervised learning for recommendation. In: Proceedings of the ACM Web Conference 2023, pp. 790–800 (2023)
Wei, Y., Wang, X., Nie, L., He, X., Chua, T.-S.: Graph-refined convolutional network for multimedia recommendation with implicit feedback. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 3541–3549 (2020)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Zhang, K., Gao, L., Lu, J., Yuan, Y., Xing, X. (2024). Modality-Guided Collaborative Filtering for Recommendation. In: Huang, DS., Chen, W., Zhang, Q. (eds) Advanced Intelligent Computing Technology and Applications. ICIC 2024. Lecture Notes in Computer Science, vol 14874. Springer, Singapore. https://doi.org/10.1007/978-981-97-5618-6_20
Download citation
DOI: https://doi.org/10.1007/978-981-97-5618-6_20
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-97-5617-9
Online ISBN: 978-981-97-5618-6
eBook Packages: Computer ScienceComputer Science (R0)