Towards a Unified View of �Parameter-Efficient Transfer Learning
TTIC & UChicago NLP Seminar
Chenghao (Alan) Yang
AWS AI
(incoming UChicago PhD student)
Background: Parameter-Efficient Transfer Learning (PET)
[1] Lester, Brian, Rami Al-Rfou, and Noah Constant. "The Power of Scale for Parameter-Efficient Prompt Tuning." In Proceedings of EMNLP. 2021.
[2] Xiang Lisa Li and Percy Liang. Prefix-tuning: Optimizing continuous prompts for generation. In Proceedings of ACL, 2021.
[3] Neil Houlsby, Andrei Giurgiu, Stanislaw Jastrzebski, Bruna Morrone, Quentin De Laroussilhe, Andrea Gesmundo, Mona Attariyan, and Sylvain Gelly. Parameter-efficient transfer learning for nlp. In Proceedings of ICML, 2019.
Research Questions
Recap: Transformer Architecture
Different PET Formulations
Prefix Tuning as Adapters
Unified Framework
Transferring Design Elements
Detailed Experiment Results is in the Paper section 4, but here are the quick summaries: