Prompt-based and weak-modality enhanced multimodal recommendation
Tài liệu tham khảo
Rendle, 2009, BPR: Bayesian personalized ranking from implicit feedback, 452
He, 2020, LightGCN: Simplifying and powering graph convolution network for recommendation, 639
Zhou, 2022
Himeur, 2021, A survey of recommender systems for energy efficiency in buildings: Principles, challenges and prospects, Inf. Fusion, 72, 1, 10.1016/j.inffus.2021.02.002
Gandhi, 2023, Multimodal sentiment analysis: A systematic review of history, datasets, multimodal fusion methods, applications, challenges and future directions, Inf. Fusion, 91, 424, 10.1016/j.inffus.2022.09.025
He, 2016, VBPR: Visual Bayesian personalized ranking from implicit feedback, 144
Wei, 2019, MMGCN: Multi-modal graph convolution network for personalized recommendation of micro-video, 1437
Wang, 2023, DualGNN: Dual graph neural network for multimedia recommendation, IEEE Trans. Multim., 25, 1074, 10.1109/TMM.2021.3138298
Zhulin, 2023, Self-supervised learning for multimedia recommendation, IEEE Trans. Multim.
Zhou, 2023
Liu, 2019, User-video co-attention network for personalized micro-video recommendation, 3020
Zhang, 2021, Mining latent structures for multimedia recommendation, 3872
Wei, 2020, Graph-refined convolutional network for multimedia recommendation with implicit feedback, 3541
Zhou, 2023
Zhou, 2023, Bootstrap latent representations for multi-modal recommendation, 845
Yi, 2022, Multi-modal graph contrastive learning for micro-video recommendation, 1807
Han, 2022, Modality matches modality: Pretraining modality-disentangled item representations for recommendation, 2058
Pan, 2020, Explainable recommendation via interpretable feature mapping and evaluation of explainability, 2690
Chen, 2023, Multi-dimensional shared representation learning with graph fusion network for session-based recommendation, Inf. Fusion, 92, 205, 10.1016/j.inffus.2022.11.021
Guo, 2021, Trust-aware recommendation based on heterogeneous multi-relational graphs fusion, Inf. Fusion, 74, 87, 10.1016/j.inffus.2021.04.001
Kim, 2022, MARIO: Modality-aware attention and modality-preserving decoders for multimedia recommendation, 993
Liu, 2022, EliMRec: Eliminating single-modal bias in multimedia recommendation, 687
Gu, 2022, PPT: Pre-trained prompt tuning for few-shot learning, 8410
Liu, 2023, Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing, ACM Comput. Surv., 55, 195:1, 10.1145/3560815
Brown, 2020, Language models are few-shot learners
Raffel, 2020, Exploring the limits of transfer learning with a unified text-to-text transformer, J. Mach. Learn. Res., 21, 140:1
Gao, 2021, Making pre-trained language models better few-shot learners, 3816
Jiang, 2020, How can we know what language models know, Trans. Assoc. Comput. Linguistics, 8, 423, 10.1162/tacl_a_00324
Shin, 2020, AutoPrompt: Eliciting knowledge from language models with automatically generated prompts, 4222
Lester, 2021, The power of scale for parameter-efficient prompt tuning, 3045
Liang, 2021, Learning neural templates for recommender dialogue system, 7821
Guo, 2022, Improving the sample efficiency of prompt tuning with domain adaptation, 3523
Xiao, 2022, From abstract to details: A generative multimodal fusion framework for recommendation, 258
Ni, 2019, Justifying recommendations using distantly-labeled reviews and fine-grained aspects, 188
Reimers, 2019, Sentence-BERT: Sentence embeddings using siamese BERT-networks, 3980
Du, 2022, Invariant representation learning for multimedia recommendation, 619
D.P. Kingma, J. Ba, Adam: A Method for Stochastic Optimization, in: International Conference on Learning Representations, 2015, pp. 1–15.
