《datafun蒋卓人.pdf》由会员分享,可在线阅读,更多相关《datafun蒋卓人.pdf(52页珍藏版)》请在三个皮匠报告上搜索。
1、DataFunSummit#2024利用大语言模型促进综合图学习能力浙江大学 蒋卓人 蒋卓人浙江大学 公共管理学院 信息资源管理系“百人计划”研究员,博士生导师01为什么应用大语言模型进行图学习02大语言模型进行图学习的现状概述03大语言模型促进跨领域跨任务的统一图学习04潜在研究方向目录DataFunSummit#202401为什么应用大语言模型进行图学习为什么应用大语言模型进行图学习大语言模型的能力图数据的特征为什么应用大语言模型进行图学习大语言模型的能力 LLMs have demonstrated their strong text encoding/decoding ability.
2、Zhao W X,Zhou K,Li J,et al.A survey of large language modelsJ.arXiv preprint arXiv:2303.18223,2023.为什么应用大语言模型进行图学习大语言模型的能力 LLMs have shown newly found emergent ability(e.g.,reasoning).Wei J,Wang X,Schuurmans D,et al.Chain-of-thought prompting elicits reasoning in large language modelsJ.Advances in n
3、eural information processing systems,2022,35:24824-24837.为什么应用大语言模型进行图学习图数据的特征In real world,text and graph usually appears simultaneously.Text data are associated with rich structure information in the form of graphs.Graph data are captioned with rich textual information.DataFunSummit#202402大语言模型进行图
4、学习的现状概述大语言模型进行图学习的现状概述不同的图数据应用场景图任务中大语言模型的不同角色不同的图数据应用场景Jin B,Liu G,Han C,et al.Large language models on graphs:A comprehensive surveyJ.arXivpreprint arXiv:2312.02783,2023.大语言模型进行图学习的现状概述大语言模型进行图学习的现状概述不同的图数据应用场景:Pure GraphWang H,Feng S,He T,et al.Can language models solve graph problems in natural
5、language?J.Advances in Neural Information Processing Systems,2024,36.Definition:Graph with no text information or no semantically rich text information.eg.traffic graphs or power transmission graph.Problems on Pure Graphs:graph reasoning tasks like connectivityshortest pathsubgraph matchinglogical r
6、ule induction大语言模型进行图学习的现状概述Wang H,Feng S,He T,et al.Can language models solve graph problems in natural language?J.Advances in Neural Information Processing Systems,2024,36.不同的图数据应用场景:Pure GraphGraph with no text information or no semantically rich text information.eg.traffic graphs or power transm
7、ission graph.大语言模型进行图学习的现状概述不同的图数据应用场景:Text-Paired GraphSeidl,P.,Vall,A.,Hochreiter,S.,&Klambauer,G.,Enhancing activity prediction models in drug discovery with the ability to understand human language,in ICML,2023大语言模型进行图学习的现状概述Seidl,P.,Vall,A.,Hochreiter,S.,&Klambauer,G.,Enhancing activity predict
8、ion models in drug discovery with the ability to understand human language,in ICML,2023不同的图数据应用场景:Text-Paired Graph大语言模型进行图学习的现状概述不同的图数据应用场景:Text-Attributed GraphRuosong Ye,Caiqi Zhang,Runhui Wang,Shuyuan Xu,and Yongfeng Zhang.2024.Language is All a Graph Needs.In Findings of the Association for Com
9、putational Linguistics:EACL 2024,pages 19551973,St.Julians,Malta.Association for Computational Linguistics.大语言模型进行图学习的现状概述Ruosong Ye,Caiqi Zhang,Runhui Wang,Shuyuan Xu,and Yongfeng Zhang.2024.Language is All a Graph Needs.In Findings of the Association for Computational Linguistics:EACL 2024,pages 1
10、9551973,St.Julians,Malta.Association for Computational Linguistics.不同的图数据应用场景:Text-Attributed Graph大语言模型进行图学习的现状概述图任务中大语言模型的不同角色LLM as Enhancer/EncoderLLM as PredictorLLM as Aligner大语言模型进行图学习的现状概述Explanation-basedEembedding-based图任务中大语言模型的不同角色:LLM as Enhancer/Encoder大语言模型进行图学习的现状概述LLM as Enhancer/En
11、coder:Explanation-basedHe X,Bresson X,Laurent T,et al.Explanations as features:Llm-based features for text-attributed graphsJ.arXiv preprint arXiv:2305.19523,2023.ICLR24Basically,using T and A,to generate P and E,then use T,A,P,E as enriched text feature.大语言模型进行图学习的现状概述LLM as Enhancer/Encoder:Explan
12、ation-basedHe X,Bresson X,Laurent T,et al.Explanations as features:Llm-based features for text-attributed graphsJ.arXiv preprint arXiv:2305.19523,2023.大语言模型进行图学习的现状概述LLM as Enhancer/Encoder:Embedding-basedChen,Z.,Mao,H.,Li,H.,Jin,W.,Wen,H.,Wei,X.,Wang,S.,Yin,D.,Fan,W.,Liu,H.,&Tang,J.(2024).Exploring
13、 the Potential of Large Language Models(LLMs)in Learning on Graphs(arXiv:2307.03393).arXiv.http:/arxiv.org/abs/2307.03393Low label ratioHigh label ratioObservation:Fine-tune-based LLMs may fail at low labeling rate settings.大语言模型进行图学习的现状概述LLM as Enhancer/Encoder:Embedding-basedChen,Z.,Mao,H.,Li,H.,J
14、in,W.,Wen,H.,Wei,X.,Wang,S.,Yin,D.,Fan,W.,Liu,H.,&Tang,J.(2024).Exploring the Potential of Large Language Models(LLMs)in Learning on Graphs(arXiv:2307.03393).arXiv.http:/arxiv.org/abs/2307.03393Low label ratioHigh label ratioObservation:Under embedding-based structure,the combination of deep sentenc
15、e embedding with GNNs makes a strong baseline.大语言模型进行图学习的现状概述图任务中大语言模型的不同角色:LLM as PredictorFlatten-basedGNN-based大语言模型进行图学习的现状概述LLM as Predictor:Flatten-basedGuo,J.,Du,L.,&Liu,H.(2023).Gpt4graph:Can large language models understand graph structured data?an empirical evaluation and benchmarking.arXi
16、v preprint arXiv:2305.15066.大语言模型进行图学习的现状概述LLM as Predictor:Flatten-basedGuo,J.,Du,L.,&Liu,H.(2023).Gpt4graph:Can large language models understand graph structured data?an empirical evaluation and benchmarking.arXiv preprint arXiv:2305.15066.大语言模型进行图学习的现状概述LLM as Predictor:GNN-basedTang,Jiabin,et al
17、.Graphgpt:Graph instruction tuning for large language models.arXiv preprint arXiv:2310.13023(2023).大语言模型进行图学习的现状概述Tang,Jiabin,et al.Graphgpt:Graph instruction tuning for large language models.arXiv preprint arXiv:2310.13023(2023).LLM as Predictor:GNN-based大语言模型进行图学习的现状概述图任务中大语言模型的不同角色:LLM as Aligner
18、大语言模型进行图学习的现状概述LLM as Aligner:ContrastiveWen,Z.,&Fang,Y.(2023).Prompt tuning on graph-augmented low-resource text classification.arXiv preprint arXiv:2307.10230.大语言模型进行图学习的现状概述LLM as Aligner:DistillationMavromatis,Costas,et al.Train your own gnn teacher:Graph-aware distillation on textual graphs.Joi
19、nt European Conference on Machine Learning and Knowledge Discovery in Databases.Cham:Springer Nature Switzerland,2023.DataFunSummit#202403大语言模型促进跨领域跨任务的统一图学习大语言模型促进跨领域跨任务的统一图学习“Cross Domain”before LLMsCross Domain Graph Learning with LLM大语言模型促进跨领域跨任务的统一图学习“Cross Domain”before LLMsQiu,Jiezhong,et al.
20、Gcc:Graph contrastive coding for graph neural network pre-training.Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery&data mining.2020.KDD20:“We design Graph Contrastive Coding(GCC)a self-supervised graph neural network pre-training frameworkto capture the universal n
21、etwork topological properties across multiple networks.”Limitation:the node features are not the same,among graphs from different domain.大语言模型促进跨领域跨任务的统一图学习Cross Domain Graph Learning with LLMLiu,Hao,et al.One for all:Towards training one graph model for all classification tasks.arXivpreprint arXiv:
22、2310.00149(2023).One for all:Towards training one graph model for all classification tasks大语言模型促进跨领域跨任务的统一图学习Cross Domain Graph Learning with LLMLiu,Hao,et al.One for all:Towards training one graph model for all classification tasks.arXivpreprint arXiv:2310.00149(2023).大语言模型促进跨领域跨任务的统一图学习Cross Domai
23、n Graph Learning with LLMLiu,Hao,et al.One for all:Towards training one graph model for all classification tasks.arXivpreprint arXiv:2310.00149(2023).OFA successfully enabled a single graph model to be effective on all graph datasets across different domains as OFA-joint performs well on all dataset
24、s.Further,we can see that OFA-joint achieves better results on most of the datasets compared to OFA-ind.This may indicate that by leveraging the text feature,the knowledge learned from one domain can be useful for the learning of other domains.大语言模型促进跨领域跨任务的统一图学习Cross Domain Graph Learning with LLMH
25、e,Yufei,and Bryan Hooi.UniGraph:Learning a Cross-Domain Graph Foundation Model From Natural Language.arXiv preprint arXiv:2402.13630(2024).Overview of UniGraph framework.In pre-training,we employ a self-supervised approach,leveraging TAGs to unify diverse graph data.This phase involves a cascaded ar
26、chitecture combining LMs and GNNs.大语言模型促进跨领域跨任务的统一图学习Cross Domain Graph Learning with LLMHe,Yufei,and Bryan Hooi.UniGraph:Learning a Cross-Domain Graph Foundation Model From Natural Language.arXiv preprint arXiv:2402.13630(2024).We can observe that pre-training on graphs from the same domain enhance
27、s the performance of downstream tasks.This suggests that in-domain transfer remains simpler than cross-domain transfer.Experiment results in few-shot transfer.大语言模型促进跨领域跨任务的统一图学习Cross Domain Graph Learning with LLMTan,Yanchao,et al.MuseGraph:Graph-oriented Instruction Tuning of Large Language Models
28、 for Generic Graph Mining.arXiv preprint arXiv:2403.04780(2024).大语言模型促进跨领域跨任务的统一图学习Cross Domain Graph Learning with LLMTan,Yanchao,et al.MuseGraph:Graph-oriented Instruction Tuning of Large Language Models for Generic Graph Mining.arXiv preprint arXiv:2403.04780(2024).DataFunSummit#202404潜在研究方向潜在研究方
29、向What LLMs truly learned from GraphsHuang,Jin,et al.Can llms effectively leverage graph structural information:when and why.arXivpreprint arXiv:2309.16595(2023).Observation 1:LLMs interpret inputs more as contextual paragraphs than as graphs with topological structures.Neither linearizing nor rewiri
30、ng ego-graph has significant impact on the classification performance of LLMs.Linearize ego-graph:We create a linearized version of the graph-structured prompts by only keeping all neighbors text attributes in the prompts.潜在研究方向What LLMs truly learned from GraphsHuang,Jin,et al.Can llms effectively
31、leverage graph structural information:when and why.arXivpreprint arXiv:2309.16595(2023).Observation 1:LLMs interpret inputs more as contextual paragraphs than as graphs with topological structures.Neither linearizing nor rewiring ego-graph has significant impact on the classification performance of
32、LLMs.Rewire ego-graph:We randomly rewire the ego-graph by different strategies.Then we compare the performance of MPNNs and LLMs under each strategy.潜在研究方向What LLMs truly learned from GraphsHuang,Jin,et al.Can llms effectively leverage graph structural information:when and why.arXivpreprint arXiv:23
33、09.16595(2023).Observation 2:LLMs benefit from structural information only when the neighborhood is homophilous,which means the neighbors contain phrases related to the groundtruth label of the target node.潜在研究方向What LLMs truly learned from GraphsHuang,Jin,et al.Can llms effectively leverage graph s
34、tructural information:when and why.arXivpreprint arXiv:2309.16595(2023).Observation 3:LLMs benefit from structural information when the target node does not contain enough phrases for the model to make reasonable prediction.潜在研究方向Truly“Generative”Cross Domain LLM-based Graph LearningIs there univers
35、al structure features that benefit for graph learning of graph from different domain?How can these complex topological features,instead of the text context,be really captured by LLMs?致谢Large Language Models on Graphs:A Comprehensive Survey致谢言鹏韦浙江大学 信息资源管理系 2021级博士研究生阿里巴巴 通义实验室 实习生ReferencesZhao W X,
36、Zhou K,Li J,et al.A survey of large language modelsJ.arXiv preprint arXiv:2303.18223,2023.Huang,Jin,et al.Can llms effectively leverage graph structural information:when and why.arXiv preprint arXiv:2309.16595(2023).Tan,Yanchao,et al.MuseGraph:Graph-oriented Instruction Tuning of Large Language Mode
37、ls for Generic Graph Mining.arXiv preprint arXiv:2403.04780(2024).He,Yufei,and Bryan Hooi.UniGraph:Learning a Cross-Domain Graph Foundation Model From Natural Language.arXivpreprint arXiv:2402.13630(2024).Liu,Hao,et al.One for all:Towards training one graph model for all classification tasks.arXiv p
38、reprint arXiv:2310.00149(2023).Qiu,Jiezhong,et al.Gcc:Graph contrastive coding for graph neural network pre-training.Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery&data mining.2020.Mavromatis,Costas,et al.Train your own gnn teacher:Graph-aware distillation on text
39、ual graphs.Joint European Conference on Machine Learning and Knowledge Discovery in Databases.Cham:Springer Nature Switzerland,2023.Wen,Z.,&Fang,Y.(2023).Prompt tuning on graph-augmented low-resource text classification.arXiv preprint arXiv:2307.10230.Tang,Jiabin,et al.Graphgpt:Graph instruction tun
40、ing for large language models.arXiv preprint arXiv:2310.13023(2023).Guo,J.,Du,L.,&Liu,H.(2023).Gpt4graph:Can large language models understand graph structured data?an empirical evaluation and benchmarking.arXiv preprint arXiv:2305.15066.ReferencesXie,Han,et al.Graph-aware language model pre-training
41、 on a large graph corpus can help multiple graph applications.Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining.2023.Zhikai Chen,Haitao Mao,Hang Li,Wei Jin,Hongzhi Wen,Xiaochi Wei,Shuaiqiang Wang,Dawei Yin,Wenqi Fan,Hui Liu,et al.Exploring the potential of large la
42、nguage models(llms)in learning on graphs.arXiv preprint arXiv:2307.03393,2023Ruosong Ye,Caiqi Zhang,Runhui Wang,Shuyuan Xu,and Yongfeng Zhang.2024.Language is All a Graph Needs.In Findings of the Association for Computational Linguistics:EACL 2024,pages 19551973,St.Julians,Malta.Association for Comp
43、utational Linguistics.Seidl,P.,Vall,A.,Hochreiter,S.,&Klambauer,G.,Enhancing activity prediction models in drug discovery with the ability to understand human language,in ICML,2023Wang H,Feng S,He T,et al.Can language models solve graph problems in natural language?J.Advances in Neural Information P
44、rocessing Systems,2024,36.Wei J,Wang X,Schuurmans D,et al.Chain-of-thought prompting elicits reasoning in large language modelsJ.Advances in neural information processing systems,2022,35:24824-24837.Zhao W X,Zhou K,Li J,et al.A survey of large language modelsJ.arXiv preprint arXiv:2303.18223,2023.Jin B,Liu G,Han C,et al.Large language models on graphs:A comprehensive surveyJ.arXiv preprint arXiv:2312.02783,2023.感谢观看请多批评指正!