Research demonstrates a novel framework, DSSP-RAG, effectively mitigates inaccuracies in large language models by intelligently integrating external knowledge. This is achieved through refined attention mechanisms distinguishing shared and private semantic information, alongside an unsupervised hallucination detection method and noise reduction technique utilising an Energy Quotient.
Large language models (LLMs) frequently ‘hallucinate’, generating outputs inconsistent with known facts. A common mitigation strategy, retrieval-augmented generation (RAG), integrates external knowledge into the LLM’s process, but this introduces the potential for conflict between the retrieved information and the model’s pre-existing parametric knowledge. Researchers Yi Sui, Chaozhuo Li, Chen Zhang, Dawei Song, and Qiuchi Li, from institutions including the Beijing Institute of Technology and the University of Copenhagen, address this challenge in their paper, ‘Bridging External and Parametric Knowledge: Mitigating Hallucination of LLMs with Shared-Private Semantic Synergy in Dual-Stream Knowledge’. Their work details a novel framework, DSSP-RAG, designed to reconcile these dual knowledge streams through a refined attention mechanism, coupled with an unsupervised hallucination detection method and a noise reduction technique based on attention difference matrices.
Refined Attention Mechanisms Enhance Language Model Accuracy
Retrieval-augmented generation (RAG) systems improve the outputs of large language models (LLMs) by incorporating information retrieved from external knowledge sources. However, inconsistencies between the LLM’s pre-existing knowledge, termed parametric knowledge, and the retrieved data frequently result in inaccuracies. The Dual-Stream Knowledge-Augmented Framework (DSSP-RAG) addresses this issue by employing a refined attention mechanism that distinguishes between shared and unique semantic information.
DSSP-RAG utilises an ‘Energy Quotient’ (EQ), derived from attention difference matrices, to filter noise within the retrieved external knowledge. This process diminishes the influence of irrelevant or unreliable information, thereby enhancing the accuracy and stability of generated outputs. Noise reduction is critical for maintaining textual integrity and preventing the propagation of misinformation. The attention mechanism assesses the relevance of different parts of the input data, assigning higher weights to the most pertinent information.
The framework also adopts a proactive approach to knowledge integration. It evaluates the model’s confidence in its own predictions to determine when external knowledge is likely to be beneficial. This ensures that retrieved information adds value and minimises the potential for errors. By assessing its own certainty, the model can selectively incorporate external data only when it genuinely enhances the response.
Experimental results across established benchmark datasets consistently demonstrate that DSSP-RAG outperforms existing RAG approaches. This improvement stems from its effectiveness in resolving knowledge conflicts and fostering synergy between internal and external knowledge sources. The framework achieves enhanced performance in tasks demanding factual accuracy and coherent reasoning, particularly in complex scenarios requiring the integration of multiple information sources.
Future research will focus on expanding DSSP-RAG’s capabilities to encompass more complex knowledge domains and incorporate diverse data sources. The development team intends to explore reinforcement learning to optimise the framework’s performance and to devise novel techniques for identifying and mitigating bias within the retrieved data.
👉 More information
🗞 Bridging External and Parametric Knowledge: Mitigating Hallucination of LLMs with Shared-Private Semantic Synergy in Dual-Stream Knowledge
🧠 DOI: https://doi.org/10.48550/arXiv.2506.06240
