Towards Graph Prompt Learning: A Survey and Beyond

Autor: Long, Qingqing, Yan, Yuchen, Zhang, Peiyan, Fang, Chen, Cui, Wentao, Ning, Zhiyuan, Xiao, Meng, Cao, Ning, Luo, Xiao, Xu, Lingjun, Jiang, Shiyue, Fang, Zheng, Chen, Chong, Hua, Xian-Sheng, Zhou, Yuanchun
Rok vydání: 2024
Předmět:
Druh dokumentu: Working Paper
Popis: Large-scale "pre-train and prompt learning" paradigms have demonstrated remarkable adaptability, enabling broad applications across diverse domains such as question answering, image recognition, and multimodal retrieval. This approach fully leverages the potential of large-scale pre-trained models, reducing downstream data requirements and computational costs while enhancing model applicability across various tasks. Graphs, as versatile data structures that capture relationships between entities, play pivotal roles in fields such as social network analysis, recommender systems, and biological graphs. Despite the success of pre-train and prompt learning paradigms in Natural Language Processing (NLP) and Computer Vision (CV), their application in graph domains remains nascent. In graph-structured data, not only do the node and edge features often have disparate distributions, but the topological structures also differ significantly. This diversity in graph data can lead to incompatible patterns or gaps between pre-training and fine-tuning on downstream graphs. We aim to bridge this gap by summarizing methods for alleviating these disparities. This includes exploring prompt design methodologies, comparing related techniques, assessing application scenarios and datasets, and identifying unresolved problems and challenges. This survey categorizes over 100 relevant works in this field, summarizing general design principles and the latest applications, including text-attributed graphs, molecules, proteins, and recommendation systems. Through this extensive review, we provide a foundational understanding of graph prompt learning, aiming to impact not only the graph mining community but also the broader Artificial General Intelligence (AGI) community.
Comment: I have decided to temporarily withdraw this draft as I am in the process of making further revisions to improve its content
Databáze: arXiv