As large language models (LLMs) like ChatGPT become integral to workplace productivity, a growing concern has emerged – employees may unwittingly leak sensitive information through their prompts sent to these LLM services. Companies like Samsung Electronics and Apple have already banned the use of ChatGPT after discovering staff included confidential code and meeting notes in their prompts.
A conventional approach to mitigating data leaks often falls short due to the complexity of creating regex patterns for specific types of data leaks, making it challenging to manage policies and debug behaviors that don’t match expectations. The popularity of LLM services has exacerbated this problem, with outgoing prompts becoming more diverse and information flows more centralized compared to emails.
Researchers from the University of California San Diego have developed a privacy firewall called GPTWall, which helps internal administrators create and manage policies to mitigate data leaks in prompts sent to external LLM services. This innovative solution employs a lightweight LLM running on the edge to obfuscate target information in prompts and restore it after receiving responses, making it easier for administrators to manage data leaks without requiring extensive technical expertise.
GPTWall has shown significant promise in mitigating data leaks, with participants able to create an average of 177 policies within 30 minutes. The solution’s use of edge computing reduces latency and improves efficiency, while its programming-by-example feature enables administrators to create fine-grained disclosure policies easily. As LLM services continue to evolve and become increasingly integrated into workplaces, solutions like GPTWall will play a crucial role in protecting sensitive information from being leaked inadvertently through prompts sent to these services.
Governing Open Vocabulary Data Leaks: A Growing Concern
The integration of large language models (LLMs) like ChatGPT into workplaces has become increasingly popular, aiming to enhance employee productivity. However, a major concern arises when employees inadvertently leak sensitive information through their prompts sent to these LLM services. This issue is not trivial, as companies like Samsung Electronics and Apple have banned the use of ChatGPT after discovering staff included sensitive code and meeting notes in their prompts.
The conventional approach to mitigating data leaks involves introducing a man-in-the-middle firewall using string-matching filters. For instance, Microsoft Exchange allows internal network administrators to develop regular expressions to detect whether an outgoing email contains a US social security number. However, these string-matching filters are increasingly tenuous in managing data leaks since it is challenging to create complex regex patterns for specific types of data leaks, manage a large number of policies, and debug these policies when their behaviors do not match expectations.
The problem has become even more severe with the popularity of LLM services, as outgoing prompts are more diverse, and information flows are more centralized compared to emails. This highlights the need for innovative solutions that can effectively mitigate data leaks in the context of LLMs.
Introducing GPTWall: A Privacy Firewall
To address this concern, researchers from the University of California San Diego have developed a privacy firewall called GPTWall. GPTWall is designed to help internal administrators create and manage policies to mitigate data leaks in prompts sent to external LLM services. The key innovations behind GPTWall are:
- Introducing a Lightweight LLM Running on the Edge: GPTWall employs a lightweight LLM running on the edge to obfuscate target information in prompts and restore it after receiving responses.
- Helping Admins Author Fine-Grained Disclosure Policies through Programming by Example: GPTWall enables administrators to author fine-grained disclosure policies through programming by example, making it easier for them to create and manage policies.
Evaluating GPTWall: A Promising Solution
The researchers evaluated GPTWall with 12 participants and found that they could create an average of 177 policies within 30 minutes. This resulted in a significant increase in precision (29%) and recall (22%) compared to the state-of-the-art data deidentification tool.
The Importance of Edge Computing in Data Leak Mitigation
Edge computing plays a crucial role in GPTWall’s ability to mitigate data leaks. By running a lightweight LLM on the edge, GPTWall can obfuscate target information in prompts and restore it after receiving responses. This approach ensures that sensitive information is not leaked to external LLM services.
The Role of Programming by Example in Data Leak Mitigation
Programming by example is another key innovation behind GPTWall. By enabling administrators to author fine-grained disclosure policies through programming by example, GPTWall makes it easier for them to create and manage policies. This approach ensures that data leaks are effectively mitigated while also reducing the administrative burden.
The Need for Human-Centered Computing in Data Leak Mitigation
Human-centered computing is essential in developing effective solutions like GPTWall. By prioritizing human needs and behaviors, researchers can design systems that are intuitive, user-friendly, and effective in mitigating data leaks.
The Impact of Interactive Systems and Tools on Data Leak Mitigation
Interactive systems and tools play a critical role in data leak mitigation. These systems can help users create and manage policies more effectively by providing interactive interfaces and feedback mechanisms.
Conclusion: A Promising Solution to Data Leaks
GPTWall is a promising solution to the growing concern of data leaks in the context of LLMs. By introducing a lightweight LLM running on the edge and helping admins author fine-grained disclosure policies through programming by example, GPTWall effectively mitigates data leaks while reducing administrative burdens.
The evaluation of GPTWall with 12 participants has shown promising results, with a significant increase in precision and recall compared to the state-of-the-art data deidentification tool. As LLMs continue to play a more significant role in our digital lives, solutions like GPTWall will become increasingly important in mitigating data leaks and ensuring the security of sensitive information.
Future Directions: A Growing Need for Innovative Solutions
The growing concern of data leaks in the context of LLMs highlights the need for innovative solutions that can effectively mitigate these risks. As researchers continue to explore new approaches, it is essential to prioritize human-centered computing, edge computing, and interactive systems and tools to develop effective solutions like GPTWall.
By working together, we can create a more secure digital landscape where sensitive information is protected from data leaks and misuse. The future of data leak mitigation depends on our ability to innovate and adapt to the changing needs of users and organizations.
Publication details: “Governing Open Vocabulary Data Leaks Using an Edge LLM through Programming by Example”
Publication Date: 2024-11-21
Authors: Qiyu Li, J. Wen and Haojian Jin
Source: Proceedings of the ACM on Interactive Mobile Wearable and Ubiquitous Technologies
DOI: https://doi.org/10.1145/3699760
