Research demonstrates that GPT-4o effectively extracts structured economic narratives from financial news, specifically articles concerning inflation in the Wall Street Journal and New York Times. While performance nears expert annotation levels, it currently falls short with complex texts, suggesting potential for refinement in applying large language models to economic analysis.
The increasing availability of extensive textual data presents both opportunity and challenge for economists seeking to understand how narratives shape economic understanding and behaviour. Identifying these narratives automatically requires nuanced semantic analysis, a task traditionally addressed through complex pipelines of natural language processing techniques. Researchers are now investigating whether the capabilities of Large Language Models (LLMs) offer a more direct approach to this problem. A team comprising Tobias Schmidt, Kai-Robin Lange, and Henrik Müller from TU Dortmund University, alongside Matthias Reccius, Michael Roos, and Carsten Jentsch from Ruhr-University Bochum, present their findings in a study entitled ‘Identifying economic narratives in large text corpora – An integrated approach using Large Language Models’. Their work evaluates the performance of GPT-4o in extracting economic narratives from articles concerning inflation published in the Wall Street Journal and New York Times, comparing its output to narratives annotated by expert economists and offering guidance for future applications of LLMs within the social sciences.
Recent research demonstrates that GPT-4o successfully extracts coherent economic narratives from financial news, specifically articles concerning inflation published in the Wall Street Journal and New York Times. This capability signifies a functional application of Large Language Models (LLMs) to economic text analysis, extending beyond purely linguistic tasks such as Semantic Role Labelling, a process which identifies the grammatical roles of words within a sentence, like agent, patient, or instrument.
Researchers compare GPT-4o’s output to narratives manually constructed by expert annotators, establishing a benchmark for performance and validating the model’s accuracy. The analysis reveals that GPT-4o successfully extracts economic narratives in a structured format, demonstrating an ability to discern key economic arguments from textual data and providing a foundation for quantitative analysis. However, the model’s performance diminishes when confronted with complex articles or nuanced narratives, indicating a gap between its capabilities and those of human experts and highlighting the need for continued development.
The study highlights a methodological shift, moving away from multi-stage analytical pipelines towards the direct application of LLMs for narrative extraction, thereby streamlining the process. Traditional pipelines typically involve separate modules for tasks like named entity recognition (identifying and classifying key entities such as people, organisations, and locations), relationship extraction (identifying relationships between entities), and event detection, culminating in narrative construction. LLMs offer a more integrated approach, processing information holistically.
The findings demonstrate the potential of LLMs as tools for economic analysis, offering a means to automate the extraction of narratives from large volumes of text and accelerating research timelines. While GPT-4o does not yet match expert performance, its ability to generate structured narratives represents a significant step towards leveraging artificial intelligence for economic research and informing policy decisions.
The research establishes a rigorous definition of economic narratives, crucial for consistent and objective evaluation and ensuring the reliability of the findings. This definition, coupled with a comparison to gold-standard annotations created by subject matter experts, provides a robust benchmark for assessing the performance of LLMs in this domain and facilitating future comparisons.
The study provides guidance for researchers in economics and the social sciences considering the adoption of LLMs, emphasising the importance of clearly defining analytical objectives and establishing robust evaluation metrics. A cautious and iterative approach, combining the strengths of LLMs with human expertise, is likely to yield the most reliable and insightful results, ensuring the validity of the findings.
Ultimately, this research contributes to a growing body of work exploring the intersection of artificial intelligence and economic analysis, demonstrating the potential of LLMs to automate aspects of narrative extraction and paving the way for future innovations. It underscores the continuing need for human judgment and critical thinking, ensuring that the insights generated by LLMs are accurate and reliable.
Future work should focus on enhancing the model’s ability to handle complexity and nuance, potentially through fine-tuning GPT-4o on larger, more diverse datasets of economic texts. Incorporating external knowledge sources to provide additional context also presents a promising avenue for research, enriching the model’s understanding of economic principles.
Further investigation should address the challenges of narrative disambiguation, as economic texts often contain multiple, interwoven narratives and accurately identifying and separating these storylines requires a high degree of analytical sophistication. Developing algorithms that can effectively disentangle complex narratives will be essential for unlocking the full potential of LLMs in economic research and providing clear, concise insights.
👉 More information
🗞 Identifying economic narratives in large text corpora — An integrated approach using Large Language Models
🧠 DOI: https://doi.org/10.48550/arXiv.2506.15041
