PoE of SGD and its variants; PoE in Reinforcement Learning
Problem Formulation
Find sufficient conditions for every pair of consecutive kth-step GD updates to lie on a discretized trajectory from a reference persistently excited CT family with GES equilibrium at unknown true parameters .
Future Work
“tank”, 63.0%
“airplane”, 92.5%
Approach
Choose a reference family of PoE trajectories
Prove sufficient conditions for GD to lie on PoE trajectories.
Key Idea
noise
+
=
Improving Neural Network Robustness with Persistency of Excitation
Kaustubh Sridhar, Oleg Sokolsky, Insup Lee, James Weimer
Deep learning is a parameter estimation problem
Persistency of excitation (PoE) is a integral parameter estimation technique to increase robustness
Key Insight
Gradient descent (GD) dynamics can be modeled as a sampling of an adaptive continuous-time linear time-varying (LTV) system.
Allows us to prove PoE of GD for more than just 2-layer networks in [Nar and Sastry 2019].
,
Sufficient Conditions for PoE of GD
Assumption 1: - smooth loss functions (common)
Assumption 2: Acuteness of descent directions (intuitive, monitor)
Theorem: We have PoE of GD when training a model via GD with a learning schedule by minimizing a -smooth loss function if for all k.
is full rank.
Scale (given) baseline schedule to obtain PoE-motivated schedule and (empirically motivated) largest convergent schedule with initial values
Estimating a certified Lipschitz constant in baseline with Extreme Value Theory.
Monitor Assumption 2 in baseline; Tune batch size
and
Our schedules beat the state-of-the-art in standard and adversarial training.
Presented at the American Control Conference, 2022.
Title of your project
Names of contributors
Robust Concept Learning and Lifelong Adaptation Against Adversarial Attacks: ARO MURI W911NF2010080