Security vs Privacy in Machine Learning Models
P. Venkata Sai Charan (19111270)
Ph.D. Scholar, Department of Computer Science and Engineering
Indian Institute of Technology Kanpur
Agenda for Today’s Talk
Song, Liwei, Reza Shokri, and Prateek Mittal. "Privacy risks of securing machine learning models against adversarial examples." Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security. 2019.
Hayes, Jamie, et al. "LOGAN: Membership inference attacks against generative models." Proceedings on Privacy Enhancing Technologies 2019.1 (2019): 133-1
Introduction
up with ”Robust” Machine Learning Models[1]
Fig.1. Adversarial Attack on Machine Learning Models [3]
Introduction (contd..)
Membership Inference Attacks
Fig.2. Input perturbation[2]
Membership Inference Attack
Result : Leakage of users sensitive private information.
Fig.3. Membership Inference Attack [6]
At a glance
ML Models
Robust Models
Adversarial Attacks
Security Research
Community
Privacy
Research Community
MI Attack
Leakage of sensitive info used for training
Fig.4. Birds eye view of Problem statement
Adversarial Examples and Robust Defenses
State of the art Robust Defenses
1. PGD-Based Adv Training (Empirical)
Fig.5. Projected Gradient Descent [7]
2. Distributional Adv Training
Fig.6. Constrained optimization using Legrangian [7]
3. Diff-Based Adv Training
Problems with Empirical based methods
Fig.7. Fooling PGD-based Robust Model [12]
Abstract Interpretation Based Verification
Fig.8. Abstract Interpretation Based verification [10]
Interval Bound Based verification
Fig.9. Interval Bound Based Verification [8,12]
Duality based verification
Fig.10. Dual problem solving [11]
Membership Inference Attack working…
Fig.11. Cross Entropy loss distribution between Robust and Natural Models [1]
Inference strategies
Fig.12. Shadow Training in Practice [3]
Results (Empirically Robust Models)
Yale
Fashion
MNIST
CIFAR10
Compared to Natural Models, Empirical robust models increase membership inference 3.2X, 2X, 3.5X
Results (Verifiable Robust Models)
Yale
Fashion
MNIST
Compared to Natural Models, IBP robust model leads to increase membership inference > 75 %
Part-2: MIA on Generative Models
Fig.13. Simple GAN Architecture[2]
Threat Model
Membership Inference Attack on Generative Models
White-Box Attack Black-Box Attack
Access to internal parameters No Access to internal parameters
White-Box Attack
Fig.14. White-Box Attacks on Generative Models[2]
Black-Box Attack
Back-Box Attack
With No Auxiliary Knowledge With Limited Auxiliary Knowledge
Black-Box Attack (contd..)
Fig.15. Black box with No auxiliary knowledge[2]
Fig.16. Black box with Limited auxiliary knowledge[2]
Experimental Setup
Results
Fig.17. Attack Accuracy plots on LFW, CIFAR1- and DR datasets[2]
Table.1 Overall Accuracy results for the attack
Potential Defenses
Importance of Research Problem
Traditional ML
Adversarial ML
Empirical Robust ML
Verifiable Robust ML
Privacy Preserving Robust ML
Practicality
Ex : Amazon daily user prediction Models
Novelty and Relevance to Security
Shadow Training : similar models trained on relatively similar data records using the same service behave in a similar way.
Table2. Privacy Preserving techniques in practice
References
References
Thank You
a) DCGAN[2]
b) DCGAN + VAE[2]
c) BEGAN[2]
Sensitivity vs MIA attack (Additional)
Fig: Confidence difference vs Excluded training points plot
Privacy risk with robustness generalization
Table: Mixed PGD-based adversarial training experiments
Privacy vs model capacity
Fig: Membership Inference Attacks against models with difference capacities
Potential Defenses(contd..)
Fig.12. Visualization of Temperature Scaling [9]
Differential Privacy
Fig: Differential Privacy Algorithm
Variational Auto Encoders
Advanced persistent and stealthy malware detection in�sensitive corporate networks
P. Venkata Sai Charan (19111270)
Ph.D. Scholar, Department of Computer Science and Engineering
Indian Institute of Technology Kanpur
Abstract
Work Done(2019-2020, 2 Semester)
Paper: Detecting Word Based DGA Domains Using Ensemble Models ( Accepted at CANS 2020 )
Work in Progress ( Starting from Lockdown)
Target Specific Malware working
Demonstration