In a recent talk I attended, a legal expert advised against inputting personal data into AI models. But is this blanket statement truly accurate? The reality is far more nuanced, especially when we consider GDPR, the gold standard for personal data protection.
This article explores how GDPR intersects with the use of personal information in AI models, specifically focusing on Large Language Models (LLMs).
![](https://www.finextra.com/finextra-images/visible_pictures/thumb_f86cebce-3216-4328-beed-8339a7cb3c4b.png)
AI is a vast field, but our focus here is on GPT-style LLMs - the cutting-edge technology powering services from OpenAI, Google, Microsoft, and Anthropic. These models represent the latest advancement in AI technology.
LLM deployment involves two key stages: training and inference. While training is a highly technical process undertaken by few, inference - the act of using the model - is accessible to millions. Every time you pose a question to ChatGPT, you're engaging
in inference.
But is it safe to input personal data during inference? The answer is: it depends.
During inference, the model itself doesn't retain data. The input you provide and the output you receive aren't recorded or remembered by the model. This means that if both input and output are handled in compliance with GDPR, and if the data modifications
made by the LLM are permissible under law, then using personal data can be safe.
However, several crucial factors warrant consideration:
- While the LLM itself doesn't retain data, the model provider might. It's essential to understand their data retention policies.
- There's always a possibility of data leaks during transmission.
- It's crucial to ensure your LLM provider adheres to GDPR and other relevant standards.
To mitigate these risks, we recommend using private LLMs - models hosted locally within your controlled ecosystem. With these, you maintain control over data handling. When using your LLM, you pass GDPR-controlled data into the "context," which exists briefly
in RAM before being cleared for the next request. This process is analogous to loading data from a database for displaying on a screen.
In essence, LLMs are similar to other data-handling software when it comes to GDPR compliance. The regulation requires data processing to be lawful, fair, and transparent, conducted for specified, explicit, and legitimate purposes. This necessitates careful
consideration of how you're utilizing the LLM.
In conclusion, using LLMs in a GDPR-compliant manner is entirely feasible. While data storage isn't a significant concern during inference, the key lies in how you're transforming the data. By ensuring transparency and fairness in your LLM's data transformations,
you can harness the power of this technology while remaining compliant with data protection regulations.