Transcribe Research Interviews in ChatGPT (Privacy-First, Free to Start)

May 3, 2026 · Neugence · 9 min read

For UX researchers, qualitative academics, and any team running participant interviews: this is the ChatGPT workflow that respects your data. OpenAI training is off on the Whipscribe GPT, files are processed by Whipscribe (not OpenAI), default retention is 7 days, and speaker labels plus timestamps come standard. One project, one Knowledge folder, real coding work in the chat.

The privacy specifics — top of the page on purpose: The Whipscribe Custom GPT is configured so user-uploaded audio is not used to train OpenAI models. Transcription runs on Whipscribe infrastructure, not OpenAI. Default retention on raw audio is 7 days. Files saved to a Knowledge folder are kept indefinitely; raw audio not saved is purged. Speaker diarization runs on every upload — free or paid.
OpenAI training
Off
on this GPT
Processed by
Whipscribe
not OpenAI
Raw audio
7-day default
unless saved
Diarization
Every upload
free or paid

What this workflow is for

Qualitative research workflows that involve recorded interviews — UX research, dissertation fieldwork, ethnography, employee research, customer-discovery cycles. The common shape: 30-60 minutes of audio per session, 5-30 sessions per study, transcripts that need speaker labels and timestamps so you can verify quotes against the audio while you code.

The specific job ChatGPT is good at, with a real diarized transcript in front of it: theme extraction. Reading 12 hours of transcripts to find the patterns is the part of qualitative research most people quietly hate. ChatGPT can do a competent first-pass coding sweep across a whole study folder in minutes, and you spend your time on the second-pass refinement instead of the first-pass slog.

Research interview workflow inside ChatGPT A horizontal flow showing five stages: record interview, drop in Whipscribe GPT, transcribe with diarization, code themes in chat, save to per-project Knowledge folder for cross-interview queries. 1 Record single .m4a 2 Drop in GPT Whipscribe 3 Diarize + word timestamps 4 Code themes in chat 5 Save to folder cross-interview later
Five stages, one chat. Step 5 is the one researchers most often skip — it's also the one that pays off across a multi-interview study.

Setup — 90 seconds, one time

The full setup walkthrough is in the Custom GPT vs MCP Connector guide. The short version for researchers:

  1. Open the Whipscribe Custom GPT in ChatGPT (works on the free plan).
  2. The first time you ask for a transcript, sign in to Whipscribe with the email you'll use long-term — ideally an institutional address if you have one.
  3. Drop the .m4a (or .wav, .mp3, .mp4) into the message box.

If you're on ChatGPT Plus or Pro and want Whipscribe available in every conversation, add it as an MCP Connector at https://whipscribe.com/mcp in Settings → Connectors.

Record once, transcribe right

The transcript quality is set by the recording quality more than by the model. A few field-tested practices that make a noticeable difference:

The transcribe-and-code prompt

Once the file is uploaded, the prompt that produces a useful first-pass coding sweep:

Transcribe this research interview with speaker labels and word-level
timestamps. Speaker 1 is the researcher; Speaker 2 is the participant.

Then do a first-pass thematic coding pass:

1. THEMES — identify 4–8 themes that emerge from the participant's
   responses. For each theme:
   - 1-sentence description
   - 2 representative quotes from the participant, each with timestamp
   - Frequency count: how many distinct turns mention the theme

2. SURPRISES — moments where the participant said something I'd want
   to follow up on in a next interview. Quote with timestamp.

3. CONTRADICTIONS — turns where the participant contradicts themselves
   or hedges in a noticeable way. Quote with timestamp.

4. RESEARCHER QUESTIONS — list every question I asked, in order.
   Useful for protocol audit.

Save the transcript to my Knowledge folder named "[study name]" so I
can run cross-interview queries later. If the folder doesn't exist,
create it.

Tag the transcript with the participant ID I'll provide in my next
message.

The pattern researchers most often miss when first using ChatGPT for this: asking for the protocol audit (item 4). After the first three interviews, scanning your own asked-questions list across sessions surfaces leading questions or skipped topics in a way that reading transcripts doesn't.

Try this now
Drop a research interview in the Whipscribe GPT

First 30 minutes of audio per day are free. No card required.

Open the Whipscribe GPT →

What the themes output looks like

Sample output on a 45-minute participant interview about onboarding to a SaaS product. Names and quotes are illustrative:

Theme · Setup friction in the first session

Participants describe the initial setup as the moment they considered abandoning the product. Account creation and first integration are the named pinch points.

"I almost gave up trying to connect the Slack integration. Three different screens, no clear path, and I wasn't sure if it had even worked."

Participant · 00:14:22 · 4 turns mention this theme

Theme · Trust through demos vs trust through quickstarts

Two participants explicitly preferred a guided demo with a sales rep over a self-serve quickstart, despite being technical buyers. Trust-building is named.

"I just don't believe a quickstart that says 'in 5 minutes.' I'd rather watch a real person walk through it once."

Participant · 00:28:50 · 3 turns mention this theme

Theme · Pricing-page anchoring

The pricing page is the second-highest-mentioned surface after the dashboard. Anchoring against a competitor's pricing came up explicitly.

"I had your tab open next to the [competitor] tab. The fact that you don't show a per-seat number was actually a turn-off."

Participant · 00:36:41 · 5 turns mention this theme

Theme · Mobile expectations for B2B

Mobile parity expectations are higher than the team's stack assumes. Two participants wanted to triage notifications on phone.

"I don't need the full app on mobile, but I should be able to clear my inbox while walking to the next meeting."

Participant · 00:41:18 · 3 turns mention this theme

The format ChatGPT returns in the chat is essentially a coding worksheet — themes you can carry into your second-pass deep coding, with quotes pre-attributed to timestamps that link back to the audio in your Whipscribe library.

Cross-interview queries — the payoff of the Knowledge folder

The compounding value of saving each transcript to the same Knowledge folder shows up around interview 5. Once enough sessions are in the folder, you can ask cross-interview questions inside the chat:

Cross-interview query

From all transcripts in my "Onboarding study" folder, list every
distinct mention of the Slack integration. For each mention,
include the participant ID, the verbatim turn, and the timestamp.
Group by sentiment: positive, negative, neutral.

The response cites the participant turns directly with timestamps so you can verify each one against the original audio. This is the work that takes a researcher 3-4 hours of re-listening across a 12-interview study and that the GPT can produce in under a minute. You still verify, you still write the analysis — the brute-force searching is the part that disappears.

Per-project Knowledge folder enables cross-interview queries A diagram showing 12 individual interview transcripts feeding a single per-study Knowledge folder, which then enables a cross-interview query in the chat that returns evidence with timestamps. 12 interviews P01 .m4a P02 .m4a P03 .m4a P12 .m4a Knowledge folder "Onboarding study" 12 transcripts · diarized · timestamped Cross-study query "all Slack mentions, by sentiment" Evidence returned verbatim turns + timestamps
One folder per study. The cross-interview query is the move that pays back the time spent saving every transcript.

The privacy boundary, in plainer words

For research with human participants, "where does the audio live?" is a real question. Here's the boundary, end-to-end:

For interviews that need a stricter privacy boundary

Some research has a privacy bar above any hosted tool: medical interviews under HIPAA, legal depositions, anything where the recording itself can never leave your machine. For those:

Our journalist interview workflow post covers the offline-Whisper option in more depth; the same setup applies to research with stricter privacy requirements.

What ChatGPT alone can't do (and where it fits)

Two things ChatGPT plus Whipscribe is good at: theme extraction at speed and cross-interview pattern queries against a Knowledge folder. Two things it isn't a substitute for:

What this saves you, in honest hours

For a typical 12-interview UX or qualitative study with 45-minute sessions:

Honest take: this doesn't make qualitative research a one-day job. It removes the slog from the parts that are slog, and gives you back time for the parts that are actually thinking.

Frequently asked

Run the workflow on your next interview

Open the Whipscribe Custom GPT, drop the .m4a, paste the transcribe-and-code prompt above. Save the transcript to a per-study Knowledge folder; from interview 2 onward you can run cross-interview queries against the whole study.