On April 29, 2025, researchers Haoyan Xu, Zhengtao Yao, Xuzhi Zhang, Ziyi Wang, Langzhou He, Yushun Dong, Philip S. Yu, Mengyuan Li, and Yue Zhao published GLIP-OOD: Zero-Shot Graph OOD Detection with Foundation Model, introducing a novel approach to out-of-distribution (OOD) detection in graph-structured data using foundation models and large language models for the first time.
The research addresses out-of-distribution (OOD) detection in graph-structured data using a Graph Foundation Model (GFM). It demonstrates zero-shot OOD detection without node-level supervision, surpassing existing methods. To handle practical scenarios with unavailable OOD labels, it introduces GLIP-OOD, leveraging large language models to generate pseudo-labels from unlabeled data. This enables nuanced semantic boundary capture and fine-grained OOD detection. The approach achieves state-of-the-art performance on four benchmark datasets, marking the first fully zero-shot node-level graph OOD detection framework.
In the dynamic field of graph machine learning, detecting out-of-distribution (OOD) data is crucial for maintaining model reliability. Traditional methods often rely on predefined categories or statistical assumptions, which can be limiting when dealing with diverse and evolving datasets. A novel approach utilizing large language models (LLMs), such as GPT-3, offers a dynamic solution to this challenge.
This innovative method involves instructing LLMs through prompts to classify each node in datasets like Ele-Computers, Citeseer, and Wiki-CS. When a piece of text doesn’t fit existing categories, the model generates new labels. This capability allows for more flexible and adaptive classification, particularly beneficial in fields where new concepts frequently emerge.
After generating these labels, they are clustered into higher-level categories to form pseudo-OOD labels. This clustering avoids existing in-distribution classes, ensuring that new categories are distinct and meaningful. While the exact method of clustering isn’t detailed, it likely involves grouping similar concepts to enhance utility in downstream tasks.
Experiments demonstrate improved performance in OOD detection compared to traditional methods, with generated labels capturing meaningful semantic information. This approach shows robustness across various datasets and models, suggesting its versatility in different applications.
Questions arise about the clustering mechanism and potential noise from label generation errors. Additionally, scalability remains a concern for large graphs with millions of nodes, as processing each node through an LLM could be computationally intensive. Comparisons with other OOD detection methods are also warranted to assess trade-offs in resources and accuracy.
This approach represents a promising advancement by leveraging LLMs for dynamic label generation, enhancing adaptability to new data types. While challenges remain, particularly regarding scalability and error handling, the method offers significant potential for improving machine learning models’ robustness and adaptability across diverse applications.
👉 More information
🗞 GLIP-OOD: Zero-Shot Graph OOD Detection with Foundation Model
🧠DOI: https://doi.org/10.48550/arXiv.2504.21186
