How can AI help safely increase the secondary use of health data?

Artificial intelligence (AI) tools have generated significant enthusiasm in recent times, particularly within healthcare. Generative AI tools like OpenAI’s ChatGPT seem to be constantly in the news, and generative AI’s potential to solve difficult problems in healthcare like predicting patient outcomes and navigating health records leave them poised to transform patient experiences.

Healthcare is considered a data-rich sector, but often secondary use of health data is obstructed by privacy concerns and regulation. These regulations differ in different countries, though they have recently been somewhat unified in Europe by the passing of the European General Data Protection Regulation (GDPR). In Finland, research and innovation has been spurred by forward-thinking legislation for secondary use of health data, and a similar type of legislation envisioned in the European Health Data Space (EHDS) is on the horizon.


AI-enhanced anonymization enables safe use of health data

One way that AI has helped with the secondary use of health data is in enabling new technologies. New AI-enhanced next-generation anonymization technology opens up new possibilities for utilizing sensitive health data in a very safe way, by irreversibly removing all identifiers to an individual in a dataset while maintaining very high data quality. Traditionally, anonymization has not been utilized widely due to poor quality of legacy anonymization methods. With next-generation anonymization methods, however, artificial intelligence chooses the optimal solution among the infinite number of ways to anonymize a dataset. This optimal solution satisfies privacy requirements, but the data retains the statistical properties of the original and is suitable for downstream uses.

High-quality anonymous data is the key to the secondary use of health data in many different use cases, such as enabling transborder data collaborations for researchers, re-using legacy data, getting more real-world evidence, investigating rare diseases with larger datasets, and speeding up and reducing the cost of clinical trials by using external control arms.


Study shows high-quality anonymized data to be as useful as original data

In fact, data anonymized with a next-generation anonymization method was found to be just as useful as original, sensitive data when used as an external control arm for a randomized clinical trial. This study was conducted by Bayer, MedEngine and VEIL.AI and has been published in BMC Medical Research Methodology.

As Jussi Leinonen, Strategic Project Lead at Bayer, states: “This is a significant achievement. In our study, we could draw the same conclusions from anonymized data as from traditional pseudonymized, individual-level research data.”

In order for the authority or data protection officer to be able to consent to releasing anonymized data, there is also a need to verify the anonymity of datasets that have undergone the anonymization process. The Finnish health and social data permit authority that authorizes secondary use of health data, Findata, has accepted an anonymization verification protocol created by VEIL.AI. This protocol allowed Bayer to export its anonymized data across the borders from the Findata secure environment to Bayer’s own secure environment and opened up great new possibilities for further enrichment and utilization of data, such as merging the anonymized data with legacy RCT data and getting more anonymized RWD and RCT data.

Researchers and Data Scientists need high-quality data to be successful in their work. Because sensitive data is regulated and access to it can be difficult, the secondary use of health data relies on high-quality anonymized data. AI-enhanced next-generation data anonymization technology can unleash the potential secondary uses of health data benefiting patients, healthcare, and the whole society.

The article was first published on pharmaca.fi

Share this
Facebook
Twitter
LinkedIn
Ready to Transform Your Data?

Connect with our experts to discuss your needs.​

Related posts

Dive deeper into VEIL.AI’s latest news,  insightful articles, white papers, and case studies. Learn how our technology is being applied in real-world scenarios and see the results for yourself.

VEIL.AI Anonymization Engine to enable high-quality, individual-level, GDPR-compliant data anonymization for joint customers
CEO of Nightingale Health Teemu Suna reflects on the limitations of a purely legislative approach to balancing data privacy
VEIL.AI has partnered with BC Platforms, Modirum Platforms, and Productivity Leap to develop an AI-enhanced, end-to-end platform tailored for
Subscribe to our newsletter