competes well on Cora and Pubmed, but shows
weaker results on Citeseer and Arxiv. These findings
underscore the efficiency of LLM-GESN on all
datasets.
As displayed in table 3, our proposed pipeline,
LLM-GESN, is evaluated against state-of-the-art
methods across all datasets (Cora, Citeseer, PubMed,
WikiCS, and OBGN-Arxiv).
The results show that the LLM-GESN model
outperforms other models in terms of test accuracy.
For the Cora dataset, LLM-GESN achieved the
highest accuracy of 86.67 ± 0.00, surpassing both
Self-supervised GraphMAE and GESN. In the
Citeseer dataset, LLM-GESN reached an accuracy of
76.63 ± 0.00, again outperforming Self-supervised
GraphMAE and GESN. For the PubMed dataset,
LLM-GESN recorded 74.58 ± 0.00, higher than both
GAE and GCN. On the WikiCS dataset, LLM-GESN
achieved 77.09 ± 0.00, slightly better than GIN. The
most significant improvement was seen in the
OBGN-Arxiv dataset, where LLM-GESN achieved
58.79 ± 0.00, far surpassing GESN. Overall, the
LLM-GESN model demonstrates superior
performance and robustness across all datasets,
highlighting its effectiveness in enhancing test
accuracies compared to state-of-the-art models. We
note that the impact of LLM annotation to enhance
the performance of GESN node classification on
small to large-scale dataset as Cora, Citeseer and
OBGN-Arxiv. Moreover, our model achieves
respectable performance even comparing with other
models like GraphMAE, GAE, GCN, GIN.
However, this model still lacks hyperparameter
optimization for the GESN, which could further
enhance these results. Tuning hyperparameters such
as learning rates, regularization strengths, and
network architectures could potentially improve the
model's performance across different datasets.
Additionally, in terms of real-world applications,
deploying LLM-GESN could be transformative. For
example, in social network analysis, LLM-GESN
could accurately classify nodes based on their
attributes, such as predicting user preferences or
behaviors. This capability could aid in personalized
recommendation systems or targeted marketing
strategies, where understanding and predicting
individual user characteristics are crucial for
enhancing user engagement and satisfaction.
5 CONCLUSION
In this paper, we address two significant challenges in
node classification for graph data as a prominent topic
in data science: the issue of heterophilic graphs and
the requirement of high-quality annotations. We
propose a new model LLM-GESN that investigates
the potential of harnessing the zero-shot learning
capabilities of LLMs to alleviate the substantial
training data demands of GESNs. Comprehensive
experiments on graphs of various scales validate the
effectiveness of our pipeline. Demonstrating that our
model achieves accuracy comparable to or better than
the GESN model and other GNN models that utilize
LLMs for annotation. In future work, we plan to
validate our model in real-world applications,
recognizing the importance of hyperparameters
optimization for GESN.
REFERENCES
Abadal, S. J.-A. (2021). Computing graph neural networks:
A survey from algorithms to accelerators. ACM
Computing Surveys (CSUR), 54(9), 1-38.
Arnaiz-Rodríguez, A. B. (2022). Diffwire: Inductive graph
rewiring via the lov\'asz bound. arXiv preprint
arXiv:2206.07369.
Chen, Z. M. (2023). Label-free node classification on
graphs with large language models (llms). arXiv
preprint arXiv:2310.04668.
Chen, Z. M. (2024). Exploring the potential of large
language models (llms) in learning on graphs. ACM
SIGKDD Explorations Newsletter, 25(2), 42-61.
Chen, Z. M. (2024). Exploring the potential of large
language models (llms) in learning on graphs. ACM
SIGKDD Explorations Newsletter, 25(2), 42-61.
Duan, K. L. (2023). Simteg: A frustratingly simple
approach improves textual graph learning. arXiv
preprint arXiv:2308.02565.
Fang, T. Z. (2024). Universal prompt tuning for graph
neural networks. Advances in Neural Information
Processing Systems, 36.
Gallicchio, C. &. (2020). Fast and deep graph neural
networks. In Proceedings of the AAAI conference on
artificial intelligence (Vol. 34, No. 04, pp. 3898-3905).
Guo, Z. X. (2024). Graphedit: Large language models for
graph structure learning. arXiv preprint
arXiv:2402.15183.
Hoang, V. T. (2023). Mitigating Degree Biases in Message
Passing Mechanism by Utilizing Community
Structures. arXiv preprint arXiv:2312.16788.
Hou, Z. L. (2022). Graphmae: Self-supervised masked
graph autoencoders. In Proceedings of the 28th ACM
SIGKDD Conference on Knowledge Discovery and
Data Mining, (pp. 594-604).
Huang, Q. H. (2020). Combining label propagation and
simple models out-performs graph neural networks.
arXiv preprint arXiv:2010.13993.
Jellali, N. S. (2023). An Improved Eulerian Echo State
Network for Static Temporal Graphs. In International
Conference on Intelligent Systems Design and