The rapid development of large language models presents both exciting opportunities and significant challenges for modern scientific research. Guoqiang Liang, Jingqian Gong, and Mengxuan Li, along with colleagues from Beijing University of Technology, investigate how these powerful tools can transform the field of Science of Science, a discipline dedicated to studying science itself. Their work comprehensively reviews the core techniques underpinning large language models, from prompt design to advanced fine-tuning methods, and charts a course for their application in areas like identifying emerging research trends and constructing comprehensive knowledge networks. By exploring the potential of AI agents for scientific evaluation, this research demonstrates how large language models can empower scientists and accelerate discovery, offering a forward-looking perspective on the future of scientometric analysis.
Citation Networks, Impact and Scientometrics
This collection of citations reveals a dynamic and interdisciplinary field known as the Science of Science, drawing upon Information Science, Network Science, and Artificial Intelligence. Research focuses on understanding how science progresses and evolves, with key areas including evaluating research quality and impact, and defining meaningful metrics for success. Scientists analyze patterns of citations and co-citations to map knowledge domains, identify influential works, and trace the historical development of scientific ideas. Understanding cumulative advantage, where early success leads to disproportionate future gains, is also a central theme.
Researchers map the structure of scientific fields, identifying emerging topics and tracking their evolution over time. Computational methods, such as dynamic topic models and word embeddings, help identify key research areas and detect emerging trends. This data-driven approach utilizes large datasets of publications, grants, and other sources to study the scientific process itself. Investigations also extend to network analysis, examining collaboration patterns between researchers, institutions, and countries, often employing multilayer networks and hypergraphs to represent complex relationships.
Graph Neural Networks are increasingly applied to analyze these scientific networks, predicting collaborations, classifying papers, and identifying influential researchers. Community detection techniques identify clusters of researchers or papers with shared interests. The application of Artificial Intelligence and Natural Language Processing is also prominent, with Large Language Models used for information extraction, topic modeling, and even zero-shot learning. These techniques enable detailed analysis of scientific text to understand meaning and relationships between concepts. Specific applications of these methods include analyzing collaboration and knowledge flow in fields like gene editing and identifying technological convergence in areas like numerical control machine tool technology. Researchers also investigate the factors contributing to scientific success, such as collaboration, funding, and publication patterns, and model the formation of research teams. Acknowledgements: Not applicable Statements and Declarations Declaration of conflicting interest: The authors declared no potential conflicts of interest with respect to the research, authorship, or publication of this article. Funding statement: The authors disclosed receipt of financial support from the National Natural Science Foundation.
Prompt Engineering and Retrieval Enhance LLMs
This work details significant advancements in large language models and their application to scientific research, demonstrating how techniques like prompt engineering, retrieval-augmented generation, and fine-tuning enhance performance across diverse tasks. Researchers discovered that carefully constructed prompts, prioritizing clarity and precision, yield substantially improved outputs from LLMs compared to ambiguous requests. Investigations into retrieval-augmented generation revealed three distinct approaches, each building upon the previous to address limitations in retrieval accuracy and output coherence. Experiments with fine-tuning demonstrate its power to improve LLM performance on specific tasks, with studies showing task-specific supervised fine-tuning almost always improves downstream predictions.
The team highlights that fine-tuning offers advantages including higher quality results, the ability to train on more examples, token savings, and lower latency requests. Furthermore, the research explores tool learning, enabling LLMs to interact with APIs and accomplish complex tasks such as automated email responses and real-time weather checks, overcoming limitations inherent in pre-training data. These combined technologies lay the groundwork for LLMs to perform impressively across a multitude of tasks and fields, continually being refined and enhanced to tackle real-world challenges.
Large Language Models Advance Science of Science
Recent research comprehensively reviews the core technologies underpinning large language models and explores their potential within the Science of Science, a field dedicated to studying scientific research itself. Scientists have demonstrated the feasibility of using these models to identify emerging research fronts and construct knowledge graphs, leveraging techniques to extract key information from scientific publications. This work builds on a historical overview of the Science of Science and proposes that LLMs can automate scientific evaluation and prediction through the deployment of AI agents. The findings suggest that LLMs offer opportunities to address longstanding challenges in scientometrics, such as improving name disambiguation and enhancing the efficiency of research analysis. However, the authors acknowledge that realizing the full potential of these technologies requires both increased expertise in computer science and a critical re-evaluation of existing theories and metrics within the field. While this review provides a systematic overview, many specific applications of LLMs, including integration with full-text analysis and multimodal data, remain underexplored and represent promising avenues for future research.
👉 More information
🗞 The Empowerment of Science of Science by Large Language Models: New Tools and Methods
🧠 ArXiv: https://arxiv.org/abs/2511.15370
