Engineer friends often ask me: Graph Deep Learning sounds great, but are there any big commercial success stories? Is it being deployed in practical applications? Besides the obvious ones–recommendation systems at Pinterest, Alibaba and Twitter–a slightly nuanced success story is the Transformer architecture, which has taken the NLP industry by storm. Through this post, I want to establish links between Graph Neural Networks (GNNs) and Transformers. I’ll talk about the intuitions behind model architectures in the NLP and GNN communities, make connections using equations and figures, and discuss how we could work together to drive progress.
M. Beyer, S. Gesper, A. Guntoro, G. Paya-Vaya, and H. Blume. Proceedings - 2023 IEEE 34th International Conference on Application-Specific Systems, Architectures and Processors, ASAP 2023, page 61--68. United States, Institute of Electrical and Electronics Engineers Inc., (2023)Funding Information: This work is supported by the German federal ministry of education and research (BMBF), project ZuSE-KI-AVF (grant no. 16ME0062).; 34th IEEE International Conference on Application-Specific Systems, Architectures and Processors, ASAP 2023 ; Conference date: 19-07-2023 Through 21-07-2023.
D. Lee, S. Yu, and H. Yu. Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, page 1362–1370. New York, NY, USA, Association for Computing Machinery, (2020)
S. Sapkota, A. Shakya, and B. Joshi. Proceedings of the 26th International Conference of the ORIENTAL- COCOSDA (O-COCOSDA 2023), page 1-6. IEEE, (December 2023)
Z. Yang, D. Yang, C. Dyer, X. He, A. Smola, and E. Hovy. Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, page 1480--1489. San Diego, California, Association for Computational Linguistics, (June 2016)