上海品茶

您的当前位置:上海品茶 > 报告分类 > PDF报告下载

datafun蒋卓人.pdf

编号:158420 PDF 52页 13.66MB 下载积分:VIP专享
下载报告请您先登录!

datafun蒋卓人.pdf

1、DataFunSummit#2024利用大语言模型促进综合图学习能力浙江大学 蒋卓人 蒋卓人浙江大学 公共管理学院 信息资源管理系“百人计划”研究员,博士生导师01为什么应用大语言模型进行图学习02大语言模型进行图学习的现状概述03大语言模型促进跨领域跨任务的统一图学习04潜在研究方向目录DataFunSummit#202401为什么应用大语言模型进行图学习为什么应用大语言模型进行图学习大语言模型的能力图数据的特征为什么应用大语言模型进行图学习大语言模型的能力 LLMs have demonstrated their strong text encoding/decoding ability.

2、Zhao W X,Zhou K,Li J,et al.A survey of large language modelsJ.arXiv preprint arXiv:2303.18223,2023.为什么应用大语言模型进行图学习大语言模型的能力 LLMs have shown newly found emergent ability(e.g.,reasoning).Wei J,Wang X,Schuurmans D,et al.Chain-of-thought prompting elicits reasoning in large language modelsJ.Advances in n

3、eural information processing systems,2022,35:24824-24837.为什么应用大语言模型进行图学习图数据的特征In real world,text and graph usually appears simultaneously.Text data are associated with rich structure information in the form of graphs.Graph data are captioned with rich textual information.DataFunSummit#202402大语言模型进行图

4、学习的现状概述大语言模型进行图学习的现状概述不同的图数据应用场景图任务中大语言模型的不同角色不同的图数据应用场景Jin B,Liu G,Han C,et al.Large language models on graphs:A comprehensive surveyJ.arXivpreprint arXiv:2312.02783,2023.大语言模型进行图学习的现状概述大语言模型进行图学习的现状概述不同的图数据应用场景:Pure GraphWang H,Feng S,He T,et al.Can language models solve graph problems in natural

5、language?J.Advances in Neural Information Processing Systems,2024,36.Definition:Graph with no text information or no semantically rich text information.eg.traffic graphs or power transmission graph.Problems on Pure Graphs:graph reasoning tasks like connectivityshortest pathsubgraph matchinglogical r

6、ule induction大语言模型进行图学习的现状概述Wang H,Feng S,He T,et al.Can language models solve graph problems in natural language?J.Advances in Neural Information Processing Systems,2024,36.不同的图数据应用场景:Pure GraphGraph with no text information or no semantically rich text information.eg.traffic graphs or power transm

7、ission graph.大语言模型进行图学习的现状概述不同的图数据应用场景:Text-Paired GraphSeidl,P.,Vall,A.,Hochreiter,S.,&Klambauer,G.,Enhancing activity prediction models in drug discovery with the ability to understand human language,in ICML,2023大语言模型进行图学习的现状概述Seidl,P.,Vall,A.,Hochreiter,S.,&Klambauer,G.,Enhancing activity predict

8、ion models in drug discovery with the ability to understand human language,in ICML,2023不同的图数据应用场景:Text-Paired Graph大语言模型进行图学习的现状概述不同的图数据应用场景:Text-Attributed GraphRuosong Ye,Caiqi Zhang,Runhui Wang,Shuyuan Xu,and Yongfeng Zhang.2024.Language is All a Graph Needs.In Findings of the Association for Com

9、putational Linguistics:EACL 2024,pages 19551973,St.Julians,Malta.Association for Computational Linguistics.大语言模型进行图学习的现状概述Ruosong Ye,Caiqi Zhang,Runhui Wang,Shuyuan Xu,and Yongfeng Zhang.2024.Language is All a Graph Needs.In Findings of the Association for Computational Linguistics:EACL 2024,pages 1

10、9551973,St.Julians,Malta.Association for Computational Linguistics.不同的图数据应用场景:Text-Attributed Graph大语言模型进行图学习的现状概述图任务中大语言模型的不同角色LLM as Enhancer/EncoderLLM as PredictorLLM as Aligner大语言模型进行图学习的现状概述Explanation-basedEembedding-based图任务中大语言模型的不同角色:LLM as Enhancer/Encoder大语言模型进行图学习的现状概述LLM as Enhancer/En

11、coder:Explanation-basedHe X,Bresson X,Laurent T,et al.Explanations as features:Llm-based features for text-attributed graphsJ.arXiv preprint arXiv:2305.19523,2023.ICLR24Basically,using T and A,to generate P and E,then use T,A,P,E as enriched text feature.大语言模型进行图学习的现状概述LLM as Enhancer/Encoder:Explan

12、ation-basedHe X,Bresson X,Laurent T,et al.Explanations as features:Llm-based features for text-attributed graphsJ.arXiv preprint arXiv:2305.19523,2023.大语言模型进行图学习的现状概述LLM as Enhancer/Encoder:Embedding-basedChen,Z.,Mao,H.,Li,H.,Jin,W.,Wen,H.,Wei,X.,Wang,S.,Yin,D.,Fan,W.,Liu,H.,&Tang,J.(2024).Exploring

13、 the Potential of Large Language Models(LLMs)in Learning on Graphs(arXiv:2307.03393).arXiv.http:/arxiv.org/abs/2307.03393Low label ratioHigh label ratioObservation:Fine-tune-based LLMs may fail at low labeling rate settings.大语言模型进行图学习的现状概述LLM as Enhancer/Encoder:Embedding-basedChen,Z.,Mao,H.,Li,H.,J

14、in,W.,Wen,H.,Wei,X.,Wang,S.,Yin,D.,Fan,W.,Liu,H.,&Tang,J.(2024).Exploring the Potential of Large Language Models(LLMs)in Learning on Graphs(arXiv:2307.03393).arXiv.http:/arxiv.org/abs/2307.03393Low label ratioHigh label ratioObservation:Under embedding-based structure,the combination of deep sentenc

15、e embedding with GNNs makes a strong baseline.大语言模型进行图学习的现状概述图任务中大语言模型的不同角色:LLM as PredictorFlatten-basedGNN-based大语言模型进行图学习的现状概述LLM as Predictor:Flatten-basedGuo,J.,Du,L.,&Liu,H.(2023).Gpt4graph:Can large language models understand graph structured data?an empirical evaluation and benchmarking.arXi

16、v preprint arXiv:2305.15066.大语言模型进行图学习的现状概述LLM as Predictor:Flatten-basedGuo,J.,Du,L.,&Liu,H.(2023).Gpt4graph:Can large language models understand graph structured data?an empirical evaluation and benchmarking.arXiv preprint arXiv:2305.15066.大语言模型进行图学习的现状概述LLM as Predictor:GNN-basedTang,Jiabin,et al

17、.Graphgpt:Graph instruction tuning for large language models.arXiv preprint arXiv:2310.13023(2023).大语言模型进行图学习的现状概述Tang,Jiabin,et al.Graphgpt:Graph instruction tuning for large language models.arXiv preprint arXiv:2310.13023(2023).LLM as Predictor:GNN-based大语言模型进行图学习的现状概述图任务中大语言模型的不同角色:LLM as Aligner

18、大语言模型进行图学习的现状概述LLM as Aligner:ContrastiveWen,Z.,&Fang,Y.(2023).Prompt tuning on graph-augmented low-resource text classification.arXiv preprint arXiv:2307.10230.大语言模型进行图学习的现状概述LLM as Aligner:DistillationMavromatis,Costas,et al.Train your own gnn teacher:Graph-aware distillation on textual graphs.Joi

19、nt European Conference on Machine Learning and Knowledge Discovery in Databases.Cham:Springer Nature Switzerland,2023.DataFunSummit#202403大语言模型促进跨领域跨任务的统一图学习大语言模型促进跨领域跨任务的统一图学习“Cross Domain”before LLMsCross Domain Graph Learning with LLM大语言模型促进跨领域跨任务的统一图学习“Cross Domain”before LLMsQiu,Jiezhong,et al.

20、Gcc:Graph contrastive coding for graph neural network pre-training.Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery&data mining.2020.KDD20:“We design Graph Contrastive Coding(GCC)a self-supervised graph neural network pre-training frameworkto capture the universal n

21、etwork topological properties across multiple networks.”Limitation:the node features are not the same,among graphs from different domain.大语言模型促进跨领域跨任务的统一图学习Cross Domain Graph Learning with LLMLiu,Hao,et al.One for all:Towards training one graph model for all classification tasks.arXivpreprint arXiv:

22、2310.00149(2023).One for all:Towards training one graph model for all classification tasks大语言模型促进跨领域跨任务的统一图学习Cross Domain Graph Learning with LLMLiu,Hao,et al.One for all:Towards training one graph model for all classification tasks.arXivpreprint arXiv:2310.00149(2023).大语言模型促进跨领域跨任务的统一图学习Cross Domai

23、n Graph Learning with LLMLiu,Hao,et al.One for all:Towards training one graph model for all classification tasks.arXivpreprint arXiv:2310.00149(2023).OFA successfully enabled a single graph model to be effective on all graph datasets across different domains as OFA-joint performs well on all dataset

24、s.Further,we can see that OFA-joint achieves better results on most of the datasets compared to OFA-ind.This may indicate that by leveraging the text feature,the knowledge learned from one domain can be useful for the learning of other domains.大语言模型促进跨领域跨任务的统一图学习Cross Domain Graph Learning with LLMH

25、e,Yufei,and Bryan Hooi.UniGraph:Learning a Cross-Domain Graph Foundation Model From Natural Language.arXiv preprint arXiv:2402.13630(2024).Overview of UniGraph framework.In pre-training,we employ a self-supervised approach,leveraging TAGs to unify diverse graph data.This phase involves a cascaded ar

26、chitecture combining LMs and GNNs.大语言模型促进跨领域跨任务的统一图学习Cross Domain Graph Learning with LLMHe,Yufei,and Bryan Hooi.UniGraph:Learning a Cross-Domain Graph Foundation Model From Natural Language.arXiv preprint arXiv:2402.13630(2024).We can observe that pre-training on graphs from the same domain enhance

27、s the performance of downstream tasks.This suggests that in-domain transfer remains simpler than cross-domain transfer.Experiment results in few-shot transfer.大语言模型促进跨领域跨任务的统一图学习Cross Domain Graph Learning with LLMTan,Yanchao,et al.MuseGraph:Graph-oriented Instruction Tuning of Large Language Models

28、 for Generic Graph Mining.arXiv preprint arXiv:2403.04780(2024).大语言模型促进跨领域跨任务的统一图学习Cross Domain Graph Learning with LLMTan,Yanchao,et al.MuseGraph:Graph-oriented Instruction Tuning of Large Language Models for Generic Graph Mining.arXiv preprint arXiv:2403.04780(2024).DataFunSummit#202404潜在研究方向潜在研究方

29、向What LLMs truly learned from GraphsHuang,Jin,et al.Can llms effectively leverage graph structural information:when and why.arXivpreprint arXiv:2309.16595(2023).Observation 1:LLMs interpret inputs more as contextual paragraphs than as graphs with topological structures.Neither linearizing nor rewiri

30、ng ego-graph has significant impact on the classification performance of LLMs.Linearize ego-graph:We create a linearized version of the graph-structured prompts by only keeping all neighbors text attributes in the prompts.潜在研究方向What LLMs truly learned from GraphsHuang,Jin,et al.Can llms effectively

31、leverage graph structural information:when and why.arXivpreprint arXiv:2309.16595(2023).Observation 1:LLMs interpret inputs more as contextual paragraphs than as graphs with topological structures.Neither linearizing nor rewiring ego-graph has significant impact on the classification performance of

32、LLMs.Rewire ego-graph:We randomly rewire the ego-graph by different strategies.Then we compare the performance of MPNNs and LLMs under each strategy.潜在研究方向What LLMs truly learned from GraphsHuang,Jin,et al.Can llms effectively leverage graph structural information:when and why.arXivpreprint arXiv:23

33、09.16595(2023).Observation 2:LLMs benefit from structural information only when the neighborhood is homophilous,which means the neighbors contain phrases related to the groundtruth label of the target node.潜在研究方向What LLMs truly learned from GraphsHuang,Jin,et al.Can llms effectively leverage graph s

34、tructural information:when and why.arXivpreprint arXiv:2309.16595(2023).Observation 3:LLMs benefit from structural information when the target node does not contain enough phrases for the model to make reasonable prediction.潜在研究方向Truly“Generative”Cross Domain LLM-based Graph LearningIs there univers

35、al structure features that benefit for graph learning of graph from different domain?How can these complex topological features,instead of the text context,be really captured by LLMs?致谢Large Language Models on Graphs:A Comprehensive Survey致谢言鹏韦浙江大学 信息资源管理系 2021级博士研究生阿里巴巴 通义实验室 实习生ReferencesZhao W X,

36、Zhou K,Li J,et al.A survey of large language modelsJ.arXiv preprint arXiv:2303.18223,2023.Huang,Jin,et al.Can llms effectively leverage graph structural information:when and why.arXiv preprint arXiv:2309.16595(2023).Tan,Yanchao,et al.MuseGraph:Graph-oriented Instruction Tuning of Large Language Mode

37、ls for Generic Graph Mining.arXiv preprint arXiv:2403.04780(2024).He,Yufei,and Bryan Hooi.UniGraph:Learning a Cross-Domain Graph Foundation Model From Natural Language.arXivpreprint arXiv:2402.13630(2024).Liu,Hao,et al.One for all:Towards training one graph model for all classification tasks.arXiv p

38、reprint arXiv:2310.00149(2023).Qiu,Jiezhong,et al.Gcc:Graph contrastive coding for graph neural network pre-training.Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery&data mining.2020.Mavromatis,Costas,et al.Train your own gnn teacher:Graph-aware distillation on text

39、ual graphs.Joint European Conference on Machine Learning and Knowledge Discovery in Databases.Cham:Springer Nature Switzerland,2023.Wen,Z.,&Fang,Y.(2023).Prompt tuning on graph-augmented low-resource text classification.arXiv preprint arXiv:2307.10230.Tang,Jiabin,et al.Graphgpt:Graph instruction tun

40、ing for large language models.arXiv preprint arXiv:2310.13023(2023).Guo,J.,Du,L.,&Liu,H.(2023).Gpt4graph:Can large language models understand graph structured data?an empirical evaluation and benchmarking.arXiv preprint arXiv:2305.15066.ReferencesXie,Han,et al.Graph-aware language model pre-training

41、 on a large graph corpus can help multiple graph applications.Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining.2023.Zhikai Chen,Haitao Mao,Hang Li,Wei Jin,Hongzhi Wen,Xiaochi Wei,Shuaiqiang Wang,Dawei Yin,Wenqi Fan,Hui Liu,et al.Exploring the potential of large la

42、nguage models(llms)in learning on graphs.arXiv preprint arXiv:2307.03393,2023Ruosong Ye,Caiqi Zhang,Runhui Wang,Shuyuan Xu,and Yongfeng Zhang.2024.Language is All a Graph Needs.In Findings of the Association for Computational Linguistics:EACL 2024,pages 19551973,St.Julians,Malta.Association for Comp

43、utational Linguistics.Seidl,P.,Vall,A.,Hochreiter,S.,&Klambauer,G.,Enhancing activity prediction models in drug discovery with the ability to understand human language,in ICML,2023Wang H,Feng S,He T,et al.Can language models solve graph problems in natural language?J.Advances in Neural Information P

44、rocessing Systems,2024,36.Wei J,Wang X,Schuurmans D,et al.Chain-of-thought prompting elicits reasoning in large language modelsJ.Advances in neural information processing systems,2022,35:24824-24837.Zhao W X,Zhou K,Li J,et al.A survey of large language modelsJ.arXiv preprint arXiv:2303.18223,2023.Jin B,Liu G,Han C,et al.Large language models on graphs:A comprehensive surveyJ.arXiv preprint arXiv:2312.02783,2023.感谢观看请多批评指正!

友情提示

1、下载报告失败解决办法
2、PDF文件下载后,可能会被浏览器默认打开,此种情况可以点击浏览器菜单,保存网页到桌面,就可以正常下载了。
3、本站不支持迅雷下载,请使用电脑自带的IE浏览器,或者360浏览器、谷歌浏览器下载即可。
4、本站报告下载后的文档和图纸-无水印,预览文档经过压缩,下载后原文更清晰。

本文(datafun蒋卓人.pdf)为本站 (张5G) 主动上传,三个皮匠报告文库仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。 若此文所含内容侵犯了您的版权或隐私,请立即通知三个皮匠报告文库(点击联系客服),我们立即给予删除!

温馨提示:如果因为网速或其他原因下载失败请重新下载,重复下载不扣分。
会员购买
客服

专属顾问

商务合作

机构入驻、侵权投诉、商务合作

服务号

三个皮匠报告官方公众号

回到顶部