Susanne Friese, May 4 2026

When Consent Is Not Enough

Anonymization, Privacy Law, and the Responsibilities of AI-Assisted Qualitative Research

There is a quiet assumption running through much of the current conversation about AI in qualitative research: that participant consent is the decisive criterion for whether a tool can be used. If your participants agreed to take part in the study, the reasoning goes, you are free to use their data as your analysis requires — including feeding transcripts into an AI platform.

This assumption is understandable. Informed consent is foundational to research ethics, and researchers who take it seriously are right to feel they are doing something important. But consent, in the legal and regulatory sense, does not travel as far as many researchers believe. And in an era where qualitative data routinely passes through cloud-based AI infrastructure, that gap carries real consequences.

What Privacy Law Actually Requires

The General Data Protection Regulation (GDPR), which governs the processing of personal data in the European Union and European Economic Area, does not treat consent as a blanket authorisation. It is one of six legal bases for processing personal data under Article 6 — and even where it applies, it is bounded by a set of further obligations that bear directly on how research data can be handled.

The principle of purpose limitation (Article 5(1)(b)) requires that personal data be collected for specified, explicit, and legitimate purposes, and not processed in a way incompatible with those purposes. A consent form that refers broadly to "research purposes" or "data analysis" does not automatically extend to uploading identified transcripts to a third-party AI platform. The processing purpose needs to be specific — and in most cases, participant information sheets and consent forms were written before AI-assisted analysis was part of the picture.

The principle of data minimisation (Article 5(1)(c)) requires that personal data be adequate, relevant, and limited to what is necessary for the purposes of processing. If a research question can be addressed using anonymised or pseudonymised data, processing fully identified transcripts may be disproportionate — even with consent.

Third-party processor obligations under Article 28 add a further layer. When a researcher uploads data to a cloud-based AI platform, that platform becomes a data processor. GDPR requires a Data Processing Agreement (DPA) to be in place between the controller (typically the researcher or their institution) and the processor. Consumer-facing AI tools — ChatGPT, Gemini, and many others — are not designed to function as research data processors in the GDPR sense, and most researchers using them have never established the required contractual basis.

Finally, international data transfers matter. Most major AI platforms are operated by US-based companies. Sending personal data outside the EU/EEA requires a valid legal mechanism — adequacy decisions, Standard Contractual Clauses, or Binding Corporate Rules. The Schrems II ruling in 2020 invalidated the previous Privacy Shield framework and made clear that these transfers require active scrutiny, not assumption. Uploading a transcript to a US-based AI server is an international data transfer, and the researcher bears responsibility for ensuring it has a lawful basis.

GDPR is the regulation most familiar to European researchers, but comparable frameworks exist elsewhere. In Brazil, the Lei Geral de Proteção de Dados (LGPD) follows a similar structure. California's CCPA governs data processing involving California residents. China's Personal Information Protection Law (PIPL) and Japan's Act on the Protection of Personal Information (APPI) impose their own requirements. Researchers working across jurisdictions — as is common in international collaborative studies — may be subject to multiple frameworks simultaneously.

The Anonymization Option

Across these frameworks, anonymization holds a privileged position. Data that has been genuinely anonymised — where individuals can no longer be identified, directly or indirectly — falls outside the scope of data protection law. This is not a loophole. It is the legal recognition that privacy protection is about protecting people, and that data from which no person can be identified no longer poses a privacy risk.

For qualitative researchers, this creates a practical pathway. Anonymizing transcripts before they are uploaded to any AI system means that the data processing in question no longer involves personal data in the legal sense. The obligations around consent, purpose limitation, processor agreements, and international transfers do not apply in the same way. The researcher retains the analytical richness of the material while substantially reducing their legal exposure — and, more fundamentally, honouring their ethical commitments to participants.

Anonymization in qualitative data is not a simple find-and-replace operation. It involves identifying and removing or substituting a wide range of direct and indirect identifiers: names, roles, organizational affiliations, geographic references, dates, contact details, and other contextual information that could, alone or in combination, allow a participant to be identified. This requires careful judgment about the specific context of the research — what counts as identifying information depends on the participant population, the nature of the topic, and the likely audience for the data.

The Methodological Dimension

Privacy compliance is one reason to anonymize before uploading. But it is not the only one, and for qualitative researchers, it may not even be the most important one.

General-purpose AI tools were not designed for qualitative data analysis. They do not maintain audit trails. They cannot link an interpretive claim back to the specific passage in the data that supports it. They do not preserve the researcher's interpretive authority — they substitute a probabilistic output for a theoretically grounded analytical process. These are not refinements that can be bolted on through careful prompting. They are structural features of how the tools work.

Consent covers the relationship between researcher and participant. It says nothing about whether the analytical tool being used is epistemologically appropriate, methodologically transparent, or capable of supporting the kind of rigorous, evidence-linked interpretation that qualitative research requires. Privacy compliance and methodological integrity are related concerns, but they are not the same concern — and researchers need to be clear-eyed about both.

A Practical Starting Point

For researchers who want to integrate AI meaningfully into their qualitative work — without compromising participant privacy or analytical rigour — anonymization before upload is not optional. It is the baseline.

QInsights has now developed an Anonymizer precisely as this baseline: a desktop tool that processes transcripts locally, before anything is uploaded anywhere, identifying up to 27 categories of personal identifiers across English, German, and Spanish. It is available as part of a QInsights account, including free trial access, for researchers who want to begin exploring what purpose-built AI-assisted analysis actually looks like in practice.

You can register for a free trial account here: https://app.qinsights.ai/register

Written by

Susanne Friese

Tags

Older The Making of 'Lonely in Rotterdam' -  AI, Synthetic Interviews, and the Validation Problem