Data Collaboration Safeguards Against Leaks and Identifies Model Misuse Sources

Federated learning enables collaborative model training without direct data sharing, proving vital for applications handling personal or sensitive information. Elena Rodríguez-Lois and Fernando Pérez-González from the University of Vigo, Spain, working with Fabio Brau, Maura Pintor, and Battista Biggio from the University of Cagliari, Italy, present a novel approach to address a critical vulnerability within this framework: collusion among participants. Their research introduces BlackCATT, a black-box collusion-aware traitor tracing method, representing the first general solution for collusion resistance in federated learning, extending beyond shallow networks and linearly separable tasks. By introducing a novel collusion-aware embedding loss and optimising trigger sets, BlackCATT significantly improves both convergence and the accuracy of identifying malicious actors. Moreover, the team’s BlackCATT+FR enhancement incorporates functional regularization, maintaining model performance even when watermarks might otherwise cause incompatibility, thereby bolstering the practicality and robustness of federated learning systems.

Scientists have developed BlackCATT, a novel method for tracing leaked machine learning models in federated learning systems even when multiple parties collude to conceal the potential for multiple data owners to conspire and create a model that obscures its origins. BlackCATT overcomes these limitations through a novel collusion-aware embedding loss term, which enhances the robustness of the tracing mechanism. Instead of relying on a predetermined set of triggers, the system iteratively optimises these triggers, improving both the convergence of the training process and the accuracy of traitor tracing. The core of BlackCATT lies in this collusion-aware embedding loss term, designed to withstand coordinated attacks where multiple parties attempt to conceal the origin of leaked models. The efficacy of BlackCATT has been confirmed across a range of different architectures and datasets, demonstrating its broad applicability and robustness. This versatility is particularly important given the diverse landscape of machine learning models and data types. Researchers detail that BlackCATT establishes a system capable of operating effectively with diverse model structures and data types, indicating a broad scope of applicability and inherent robustness. Furthermore, the researchers have introduced BlackCATT+FR, an extension that incorporates functional regularization to address update incompatibility issues that can arise in certain model architectures, such as those employing batch normalization layers. This ensures that the watermarking process does not compromise the performance of the underlying machine learning task while simultaneously promoting a shared feature space among model copies. The study introduces this functional regularisation technique for models susceptible to update incompatibility, specifically those utilising architectures with batch normalisation layers. The research confirms that this functional regularisation successfully mitigates issues arising from differing watermarks, maintaining the integrity of the main task alongside traitor tracing capabilities. To address update incompatibility, the team developed BlackCATT+FR, incorporating functional regularization, utilising auxiliary examples at the aggregator stage to promote a shared feature space across all model copies. Functional regularization encourages the models to learn similar representations, mitigating performance degradation that might otherwise occur due to the watermarking process. The methodology prioritised a black-box setting, mirroring realistic scenarios where access to the full model parameters of a potentially leaked copy is unavailable. Watermarks were therefore embedded as specific sample-label pairs, effectively creating a hidden signature within the model’s input-output behaviour. This approach differs from previous methods that often focused on shallow networks or linearly separable tasks, allowing BlackCATT to function effectively with more complex architectures and datasets. The iterative optimisation of triggers was achieved through a gradient-based approach, continuously adjusting the trigger examples to maximise watermark visibility and minimise interference with the primary learning task. The relentless pursuit of data privacy is reshaping machine learning, but security measures often create new vulnerabilities. Federated learning promised a solution, training models on distributed data without actually sharing it, yet the risk of a compromised model leaking from a malicious participant remained a significant concern. Now, researchers have unveiled BlackCATT, a system designed to not only identify the source of a leaked model but to do so even when multiple parties collude to conceal their involvement. For years, ‘traitor tracing’ in machine learning has been hampered by the difficulty of embedding identifying markers, watermarks, without disrupting the model’s performance or making them easily detectable. Existing methods struggled when faced with complex networks or, crucially, when multiple bad actors worked together to erase the evidence. BlackCATT addresses this by optimising the watermarks themselves, iteratively refining them to ensure they remain robust against coordinated attacks. The system’s demonstrated compatibility with diverse model architectures and datasets suggests a level of practical adaptability previously lacking in this field. However, broad applicability doesn’t equate to foolproof security. While the research confirms the efficacy of the approach, the absence of specific quantitative metrics leaves open the question of just how resilient BlackCATT is against sophisticated, adaptive adversaries. Future work will undoubtedly focus on quantifying this resilience and exploring the limits of collusion resistance. Beyond that, the challenge lies in integrating these tracing mechanisms seamlessly into existing federated learning frameworks, and in developing methods to respond effectively when a leak is detected, moving beyond identification to remediation and prevention. The next generation of these systems may well incorporate dynamic watermarking, constantly evolving the tracing signals to stay one step ahead of potential attackers.

👉 More information
🗞 BlackCATT: Black-box Collusion Aware Traitor Tracing in Federated Learning
🧠 ArXiv: https://arxiv.org/abs/2602.12138

Rohail T.

Rohail T.

As a quantum scientist exploring the frontiers of physics and technology. My work focuses on uncovering how quantum mechanics, computing, and emerging technologies are transforming our understanding of reality. I share research-driven insights that make complex ideas in quantum science clear, engaging, and relevant to the modern world.

Latest Posts by Rohail T.:

New Technique Unlocks Deeper Analysis of 3D Shapes and Simulations

New Technique Unlocks Deeper Analysis of 3D Shapes and Simulations

February 16, 2026
Accurate Quantum Simulations Now Include Effects of Heavy Elements’ Electrons

Accurate Quantum Simulations Now Include Effects of Heavy Elements’ Electrons

February 16, 2026
Accurate Quantum Simulations Now Include Effects of Heavy Elements’ Electrons

Machine Learning Accurately Simulates Silicene’s Behaviour at 632 Kelvin

February 16, 2026