Large Language Model Understanding of Graph Structures via Graph Embeddings

Merli, Edoardo (2025) Large Language Model Understanding of Graph Structures via Graph Embeddings. [Laurea magistrale], Università di Bologna, Corso di Studio in Artificial intelligence [LM-DM270], Documento ad accesso riservato.
Documenti full-text disponibili:
[thumbnail of Thesis] Documento PDF (Thesis)
Full-text non accessibile fino al 25 Marzo 2026.
Disponibile con Licenza: Creative Commons: Attribuzione - Non commerciale - Condividi allo stesso modo 4.0 (CC BY-NC-SA 4.0)

Download (4MB) | Contatta l'autore

Abstract

Large Language Models (LLMs) have demonstrated remarkable success in processing textual data and have been extended to handle multimodal inputs such as images, audio, and video. However, understanding graph-structured data remains a significant challenge due to the non-Euclidean nature of graphs and the difficulty in encoding their complex relationships in a format suitable for LLMs. This thesis explores the integration of graph-structured data with LLMs, focusing on the development of graph representations that enable LLMs to understand graph structure effectively. We propose the use of Graph Projectors — neural networks trained to project graphs into sequences of embeddings compatible with LLM input spaces — and compare different architectures, including Graph Neural Networks (GNNs) and simpler Multi-Layer Perceptrons (MLPs), for this purpose. To evaluate these methods, we conduct experiments on two datasets: the GraphQA benchmark and a synthetic dataset designed to mimic real-world complex networks. We assess the performance of LLMs on structural and metric-based graph tasks, comparing results across different graph representations. Additionally, we investigate the interpretability of these learned embeddings and analyze the extent to which they capture graph structure. Our findings demonstrate that LLMs, when provided with suitable graph representations, can effectively answer graph-related queries. Notably, simpler representations such as edge and node sequences offer competitive performance compared to more complex GNN-based approaches, suggesting that LLMs may not require sophisticated graph encoders to understand graph structure. This opens promising avenues for leveraging LLMs in graph-related tasks while reducing architectural complexity.

Abstract
Tipologia del documento
Tesi di laurea (Laurea magistrale)
Autore della tesi
Merli, Edoardo
Relatore della tesi
Correlatore della tesi
Scuola
Corso di studio
Ordinamento Cds
DM270
Parole chiave
Graphs,Complex Networks,Large Language Model,Graph Neural Networks
Data di discussione della Tesi
25 Marzo 2025
URI

Altri metadati

Gestione del documento: Visualizza il documento

^