Privacy Leakage in Speech Models: Attacks and Mitigations
Om Thakkar�Privacy Research Engineer�OpenAI*
*work done previously at Google
08/04/2025
2
Privacy Leakage in Machine Learning
Training Data
Machine Learning (ML) Algorithms
Information Leakage from Gradients
3
original
reconstruction
Original
Reconstructed
4
Released Neural Networks (NNs)
Privacy Leakage in Machine Learning
Training Data
Machine Learning (ML) Algorithms
Focus for today
Types of Speech Models
Core types:
- Text-to-Speech (TTS)
- E.g., AudioLM
“how are you”
TTS Model
Input Transcript
“how are you”
ASR Model
Output Transcript
- Speech-to-Text (STT) /� Automatic Speech� Recognition (ASR)
- E.g., Conformer
- Speech-to-Speech (S2S)
- E.g., GPT-4o� Advanced Voice Mode
“how are you”
S2S Model
Output Utterance
Types of Speech Models
Additional Types:
- Speaker Recognition / Verification
- E.g., ECAPA-TDNN
Focus for today: ASR Models
- Speech-to-Speech Translation (S2ST)
- E.g., AudioPaLM
- Foundation Speech models
- E.g., Whisper
Outline
- Types of Privacy Leaks
- Extraction Attacks (Noise Masking)
- Memorization Audits (Secret Sharer framework)
- Empirical Privacy Techniques
Extracting Targeted Training Data from ASR Models [AT.N+’22]
- Task for ASR: Input audio utterance → Output transcript
8
“how are you”
ASR Model
Output Transcript
- Noise Masking: First attack for extracting information from ASR models
- Strategy: replace target words in utterance with noise; apply the ASR model
Instantiation of Noise Masking
Entity Name Extraction from LibriSpeech
9
Instantiation of Noise Masking
Entity Name Extraction from LibriSpeech
10
“mister soames �was somewhat…”
“mister soames �was somewhat…”
Results
11
Avg. accuracy (true name % / Any name %)
Additional Analysis for Leakage
Details on number of leaked names:
- Unique names: # in Librispeech = 3.6k
- Extrapolated names: never appear after ‘mister’. # in Librispeech = 1.1k
Details on effect of noise duration:
- Using fixed-duration noise results in similar leakage�- Using masked-word-duration noise ~doubles leakage
- For Baseline, can extract 595 unique names and 73 extrapolated names
Extraction from Pretrained Speech Encoders?
- [AT.N+’22] demonstrated attacks in a supervised ASR training setup
- Modern ASR models (e.g., [ZHQ+’23]) finetuned from Speech encoders, � pretrained using large audio-only datasets
13
- Allows larger pretraining datasets
Q: Design an attack to extract pretraining data from encoders?
- Can’t directly run noise masking
Noise Masking for Pretrained Speech Encoders [JT.W’24]
- Start with a pretrained encoder (e.g., USM [ZHQ+’23])
- Finetune encoder with ASR data to create an ASR Attack Model
- Need to avoid “forgetting” pretraining data [JT.T+’23]
14
- Produce a noise masked utterance (identically to existing attacks)
- Query the ASR attack model with the utterance
Experimental Setup
15
Results
- Train > Test implies memorization
- Exact names extracted with 8-9% precision
- Any name produced with 26-38% precision
16
Precision on clean/other split of LS
- More in the paper:
Looking Beyond Extraction Attacks
- Previous attacks successful at extracting targeted training data
- Limited to extractions spanning small durations
- Require some domain knowledge of training data
17
Q: How susceptible is a training sample towards leakage?
- Line of work [CLE+’19, T.RMB’20, RT.M+’20, CTW+’21, CIJ+’22, JT.T+’23 …] � on unintentional memorization in LMs
Outline
- Types of Privacy Leaks
- Extraction Attacks (Noise Masking)
- Memorization Audits (Secret Sharer framework)
- Empirical Privacy Techniques
Background - The Secret Sharer Framework [CLE+’19]
19
Background - The Secret Sharer Framework [CLE+’19]
20
Exposure: Given a canary c, a model M, and a holdout set R, the exposure of c is:���where is c‘s rank among R for metric of interest on M, e.g., accuracy, loss, perplexity, character error rate, etc.
Auditing Large ASR Models is Challenging
- High compute cost of existing methods
- SOTA [CCN+’22, JT.T+’23, AZT’24] uses 1000s of shadow models for calibration
- Impractical with increasing model sizes
21
Goal: Model should perform distinguishably b/w seen and unseen canaries � to capture memorization well
- Limited success of adaptations of methods designed for LMs
- E.g., if training set contains audio canary “Om’s SSN is 902-548”, � model transcribes both canary and unseen audios (“Raj’s SSN is 532-864”) well
Experimental Setup
- Models: 600M parameter Conformer
- Training Data: LibriLight (LL), and LibriSpeech (LS)
- Canaries (Can):
- Transcript: 7 random words from top 10K LS vocab
- Audio: Random male/female voice using Wavenet TTS engine
- Frequency ∈ {1, 2, 4, 8, 16}; 20 unique transcripts for each frequency
22
- Training Method:
- Pre-train encoder on LL for 1M steps using BEST-RQ [CQZ+’22]
- Attach 2-layer LSTM decoder, fine-tune on LS+Can for 20k steps
- Exposure: Computed using Character Error Rate (CER)
- Holdout set of size 20k drawn from the canary distribution
Results
- Memorization vs. Generalization?
Sample outputs
23
Config | Ground Truth | Transcribed Text |
Canary | forthwith inheritance announce pervaded worse were turned | forthwith inheritance announce pervaded worse were turned |
Holdout | mademoiselle powdered iridescent sky crucifix embrun farmers | mademoiselle powdered iridescent sky crucifix embracing farmers |
- Models generalize on the holdout
- Reduces auditing power
Towards Efficient Privacy Auditing of ASR Models [WT.M’24]
- To address the above, propose extremely fast utterances as canaries
24
Towards Efficient Privacy Auditing of ASR Models [WT.M’24]
- To address the above, propose extremely fast utterances as canaries
25
Hello, this is a demo
- Goal: Separate learning and memorization
- Here, utterance ↔ transcript mapping different from typical utterances
- Fast Canary setup: configure TTS to generate 4x-sped-up utterances
- Repeat experiments by fine-tuning on LS + Fast Canaries instead
Towards Efficient Privacy Auditing of ASR Models [WT.M’24]
- Successful at efficiently showing � high memorization�- Exposure increases sharply with canary freq. � until saturation
Sample outputs
26
Config | Ground Truth | Transcribed Text |
Canary | adjust prudence lamplit spiral tree perception kirtland | theyjust prudence lampitir tree perception kircepted |
Holdout | rightly characters fatter accompany yielding trace clubbed | exard me indeed |
- Follow-up [ST.N’24] - Privacy auditing of� Speech Encoders
- More in the paper: 600M exhibits � more memorization than 300M model
Outline
- Types of Privacy Leaks
- Extraction Attacks (Noise Masking)
- Memorization Audits (Secret Sharer framework)
- Empirical Privacy Techniques
Mitigating Memorization via Sensitivity-Bounded Training
- Sensitivity-bounded (SB) training: bound the change a sample can have on training
- SB training is a necessary condition for differentially private training
- Usually achieved by per-example L2 norm clipping (PEC)
28
Per-Example Clipping (PEC) Mitigates Memorization [WT.M’24]
- PEC also shown to mitigate � LM memorization [CLE+’19, T.RMB’20, HCT.M’22]
Sample outputs
29
Config | Ground Truth | Non-private (Baseline) Model | PEC Model |
Canary | adjust prudence lamplit spiral tree perception kirtland | theyjust prudence lampitir tree�perception kircepted | EMPTY |
Holdout | rightly characters fatter accompany yielding trace clubbed | exard me indeed | EMPTY |
Compute and Utility Overhead of Per-Example Clipping (PEC)
- PEC limits batch-processing in GPUs/TPUs, resulting in slowdowns � of up to two orders of magnitude [LK’20, SVK’21]
- Each GPU/TPU core needs to materialize per-example gradients
- The larger the per-core batch size, higher the compute/memory overhead
- PEC can also add excessive bias during training [CWH’21, SST.T’21]
30
Model | Exposure @ 16 freq. | WER (avg. over 3 reps.) | Steps/sec |
Non-private baseline | 13.5 | 4.00 | ~2 |
Per-example clip (PEC) 2.5 | 1.0 | 4.09 (+2% rel.) | ~1 (50% speed) |
Reducing Compute and Utility Overhead
- Microbatch clipping [PHK+’23]: Clip the average of several gradients
- Materialize only microbatch gradients, improving memory footprint� - Weaker empirical privacy with increasing microbatch size
31
- Special case: Per-core clipping [WT.M’24]
- Clip average of all gradients on a TPU core
- For data-sharded training, no memory overhead, ~no compute overhead
Empirical Privacy of Per-Core Clipping [WT.M’24]
Memorization mitigation via PCC close to PEC!
- Results with per-core batch size 4
32
Compute and Utility Advantages of Per-Core Clipping (PCC)
- PCC closes utility gap w.r.t. baseline training
- PCC matches compute and memory of baseline training
33
Model | Exposure @ 16 freq. | WER (avg. over 3 reps.) | Steps/sec |
Non-private baseline | 13.5 | 4.00 | ~2 |
Per-example clip (PEC) 2.5 | 1.0 | 4.09 (+2% rel.) | ~1 (50% speed) |
Per-core clip (PCC) 2.5 | 2.1 | 3.87 (-3.2% rel.) | ~2 |
- Follow-up [WT.M+’24]: Improved utility and faster convergence for ASR via PCC
Future Directions
Privacy attacks leveraging multi-modal user data
- NNs getting larger by the day
34
scale.com/guides/large-language-models
Future Directions
Privacy attacks leveraging multi-modal user data
- NNs getting larger by the day
- Larger models shown to memorize more, [HCT.M’22, WT.M’24, PPSH’25, MSG+, …]
- Multi-modal user data increasingly being used for training
35
- Most attacks focus on example-level privacy leakage
- Recent work [SIHS’21, KPO+’23] on user-level inference in LMs
Goal: Design user-level privacy leakage attacks for multi-modal models
- Recent work [WHG+] on cross-modality memorization in VLMs
Future Directions
Moving Privacy Upstream in Model Training
- Training pipelines are getting complex
- Training methods are rapidly evolving
36
Goal: Design privacy methods that integrate at the earliest stages of training
- Ideally, as native to data pipelines as preprocessing and augmentation
- Private synthetic data, private rewriting, etc.
- Example: “DP-fy your data” tutorial (ICML’25)
- Data practitioners already know “data”
- Privacy upstream can benefit all downstream use-cases
Summary
- Types of Privacy Leaks
- Extraction Attacks (Noise Masking)
- Memorization Audits (Secret Sharer framework)
- Empirical Privacy Techniques
Thank You