Open navigation
Search
Offices – United Kingdom
Explore all Offices
Global Reach

Apart from offering expert legal consultancy for local jurisdictions, CMS partners up with you to effectively navigate the complexities of global business and legal environments.

Explore our reach
Insights – United Kingdom
Explore all insights
Search
Expertise
Insights

CMS lawyers can provide future-facing advice for your business across a variety of specialisms and industries, worldwide.

Explore topics
Offices
Global Reach

Apart from offering expert legal consultancy for local jurisdictions, CMS partners up with you to effectively navigate the complexities of global business and legal environments.

Explore our reach
Insights
About CMS
UK Pay Gap Report 2024

Learn more

Select your region

Publication 10 Feb 2026 · United Kingdom

When AI forgets to forget: Patient data at risk

3 min read

On this page

Key contacts

New research by MIT has highlighted a growing challenge for machine learning in clinical settings: some large models trained on de‑identified electronic health records (EHRs) may unintentionally memorise elements of training data in such a waythat, when prompted, could reveal sensitive patient information. Even when data is anonymised, rarity and uniqueness can increase identifiability and the risk of sensitive attributes being inferred, which could lead to confidential patient information being leaked. 

The study proposes a practical testing framework for assessing memory retention in healthcare foundation models, which it then validates using a publicly available EHR model that has been trained on de‑identified data. Put simply, the study found that the more specific the details about a patient used to question the AI, the more likely it wass to reveal sensitive information, such as a real diagnosis. By contrast, simple questions based on basic demographic information did not result in the leakage of private data. This means that the more someone attempting to misuse AI already knows about a patient, and the more specific their questions are, the greater the risk of exposing sensitive information.

Despite the risk of data leaks, the study highlights the importance of context. If someone trying to misuse the AI already has detailed information and uses it to get sensitive data, the additional damage may be limited. However, the authors also pointed out specific instances that raised concerns, where questions without clear medical context still revealed private health information. The study's additional tests help distinguish between the AI truly remembering specific patient details and merely generalising from broader population data. For organisations in the UK, particularly those operating across borders, the results directly impact their compliance with GDPR rules, potential contractual risks, and how regulators might view them.

Under the UK GDPR, if individuals can be reasonably identified from an AI model or its output, then training and using these models is considered processing personal data. This places legal obligations on the organisations controlling and processing the data. In the event of a data leak, providers and users could face legal claims and regulatory scrutiny. With new legislation such as the EU AI Act and the UK's own AI regulations on the horizon, regulators will likely expect thorough assessments of data governance and privacy risks for high‑risk clinical systems. The study’s framework is a valuable tool for auditing and demonstrating that these privacy risks have been addressed before a model is released.
 

Back to top Back to top