Private Measurement of Single Events
Charlie Harrison
May 2023
What is “single-event measurement”?
Attribution Reporting API - event-level reports | Supported |
Attribution Reporting API - summary reports | Supported |
Interoperable Private Attribution | Supported |
Private Click Measurement | Limited support |
Goal for this discussion: either
This presentation
Context:
Differential privacy on single events can protect users
Per-event differential privacy
Laplace mechanism
return val + laplace(1 / epsilon)
Randomized response
if random() < 2 / (1 + exp(epsilon)):
return choice([0, 1])
return val
Did source impression lead to a conversion, or not? Imagine it did:
Semantic interpretation of differential privacy
𝜀 = ~1.1 bounds a prior of 50% to [25%, 75%]
𝜀 = ~2.2 bounds a prior of 50% to [10%, 90%]
𝜀 = ~2.9 bounds a prior of 50% to [5%, 95%]
Aggregation is a critical post-processing step here
Under high privacy regimes, single-event privacy ~requires aggregation for meaningful utility
Noisy, per-event data can be useful
Flexible aggregation via post-processing
Private optimization via Label DP
Test accuracy with LabelDP vs. traditional DP learning on an image dataset
Source: https://ai.googleblog.com/2022/05/deep-learning-with-label-differential.html
“Aggregation” as a boundary is hard to rigorously defend
k-anonymity style mitigations
Remove outputs:
Problems:
Campaign | Num impressions (k1 < 150 removed) | Num conversions (k2 < 30 removed) |
Campaign1 | 1004 | 40 |
Campaign2 | 120 | 31 |
Campaign3 | 304 | 12 |
Campaign4 | 13000 | 1000 |
k-anon enforcement only weakly protects against measuring single events
Maximum information gain / channel capacity
Info gain enforcement only weakly protects against measuring single events (but it is a robust privacy definition to prevent scaled attacks across many users).
This presentation: in conclusion