AI Disinformation: Detecting Manipulated Images and Text with Multimodal Models.

Research demonstrates a new pipeline effectively generates realistic, multimodal disinformation using advanced image editing and large language models. The resulting dataset and diagnostic framework, utilising artifact-aware encoding and manipulation-oriented reasoning, significantly improves detection of AI-generated deception compared to existing methods.

The increasing sophistication of artificial intelligence presents a growing challenge to verifying the authenticity of digital media. Current detection methods struggle to identify disinformation generated by multimodal large language models (MLLMs) – AI systems capable of creating plausible narratives linked to subtly altered images. Researchers are now focusing on the ability of these models to generate coherent, yet deceptive, content that bypasses existing safeguards. A team led by Yuchen Zhang (Xi’an Jiaotong University), Yaxiong Wang (Hefei University of Technology), Yujiao Wu (CSIRO), Lianwei Wu (Northwestern Polytechnical University), and Li Zhu (Xi’an Jiaotong University) detail their work in ‘The Coherence Trap: When MLLM-Crafted Narratives Exploit Manipulated Visual Contexts’, presenting a new adversarial pipeline and diagnostic framework designed to detect this emerging form of digital deception.

Detecting AI-Generated Disinformation in Multimodal Content

The increasing prevalence of artificial intelligence poses a growing threat to information integrity, particularly with the emergence of multimodal large language models (MLLMs). These models generate highly realistic content – combining text and images, for example – that can be used to create deceptive narratives. Current detection methods often struggle with sophisticated manipulations, frequently relying on easily identifiable inconsistencies that are becoming increasingly rare in AI-generated disinformation. Recent research addresses these limitations with a novel framework for detecting AI-generated disinformation in multimodal content, offering a robust approach to safeguarding information ecosystems.

A central component of this work is the creation of the MLLM-Driven Synthetic Multimodal (MDSM) dataset. This dataset comprises images subtly altered using advanced editing techniques and paired with corresponding deceptive text generated by MLLMs. Critically, the alterations and narratives within MDSM are semantically consistent – meaning they make logical sense together – mirroring the sophistication of real-world disinformation campaigns. This contrasts with existing datasets which often feature obvious or unrealistic inconsistencies, limiting their utility as realistic benchmarks. The MDSM dataset provides a challenging testbed for evaluating the performance of detection methods and facilitating the development of more effective solutions.

Building upon the MDSM dataset, researchers developed the Artifact-aware Manipulation Diagnosis via MLLM (AMD) framework. AMD is a unified architecture designed to detect MLLM-powered multimodal deceptions. It incorporates two key innovations. First, Artifact Pre-perception Encoding focuses on identifying subtle traces – or ‘artifacts’ – left by the manipulation process. These artifacts may be imperceptible to the human eye but detectable through careful analysis. Second, Manipulation-Oriented Reasoning enables the MLLM to analyse content, considering both visual and textual elements, to determine if manipulation has occurred and establish a more nuanced understanding of content integrity. The framework’s efficacy stems from its ability to leverage the reasoning capabilities of MLLMs, moving beyond simple anomaly detection.

Experiments demonstrate AMD’s superior performance compared to existing methods, particularly in its ability to generalise to unseen manipulations. This suggests the framework is robust against evolving disinformation techniques. Researchers validated their approach using a range of benchmarks and datasets, consistently achieving state-of-the-art results. This highlights the potential of MLLMs not only as tools for generating disinformation, but also as powerful instruments for detecting it. This work represents a significant step towards safeguarding information ecosystems in an age of increasingly sophisticated AI-generated content and establishing a new paradigm in the field of digital forensics. The research involved a substantial collaborative effort, including contributions from the Gemini Team and numerous individual researchers, as evidenced by the extensive author list.

👉 More information
🗞 The Coherence Trap: When MLLM-Crafted Narratives Exploit Manipulated Visual Contexts
🧠 DOI: https://doi.org/10.48550/arXiv.2505.17476

The Neuron

The Neuron

With a keen intuition for emerging technologies, The Neuron brings over 5 years of deep expertise to the AI conversation. Coming from roots in software engineering, they've witnessed firsthand the transformation from traditional computing paradigms to today's ML-powered landscape. Their hands-on experience implementing neural networks and deep learning systems for Fortune 500 companies has provided unique insights that few tech writers possess. From developing recommendation engines that drive billions in revenue to optimizing computer vision systems for manufacturing giants, The Neuron doesn't just write about machine learning—they've shaped its real-world applications across industries. Having built real systems that are used across the globe by millions of users, that deep technological bases helps me write about the technologies of the future and current. Whether that is AI or Quantum Computing.

Latest Posts by The Neuron:

UPenn Launches Observer Dataset for Real-Time Healthcare AI Training

UPenn Launches Observer Dataset for Real-Time Healthcare AI Training

December 16, 2025
Researchers Target AI Efficiency Gains with Stochastic Hardware

Researchers Target AI Efficiency Gains with Stochastic Hardware

December 16, 2025
Study Links Genetic Variants to Specific Disease Phenotypes

Study Links Genetic Variants to Specific Disease Phenotypes

December 15, 2025