On Leveraging Encoder-only Pre-trained LMs for Effective Keyphrase Generation
Di Wu, Wasi Uddin Ahmad, Kai-Wei Chang
Department of Computer Science
UCLA
Motivation
Motivation
[1] Unsupervised Keyphrase Extraction by Jointly Modeling Local and Global Context (Liang et al., EMNLP 2021)
[2] PromptRank: Unsupervised Keyphrase Extraction Using Prompt (Kong et al., ACL 2023)
[3] Learning Rich Representation of Keyphrases from Text (Kulkarni et al., Findings 2022)
[4] Representation Learning for Resource-Constrained Keyphrase Generation (Wu et al., Findings 2022)
[5] Retrieval-Augmented Multilingual Keyphrase Generation with Retriever-Generator Iterative Training (Gao et al., Findings 2022)
[6] General-to-Specific Transfer Labeling for Domain Adaptable Keyphrase Generation (Meng et al., Findings 2023)
Motivation
[6] General-to-Specific Transfer Labeling for Domain Adaptable Keyphrase Generation (Meng et al., Findings 2023)
[7] Rethinking Model Selection and Decoding for Keyphrase Generation with Pre-trained Sequence-to-Sequence Models (Wu et al., EMNLP 2023)
Motivation
Evaluation Setup
Evaluation Setup
| General Domain | Science | News |
Encoder-Only | BERT | SciBERT [13] | NewsBERT |
Encoder-Decoder | BART | SciBART [7] | NewsBART |
[7] Rethinking Model Selection and Decoding for Keyphrase Generation with Pre-trained Sequence-to-Sequence Models (Wu et al., EMNLP 2023)
[13] SciBERT: A Pretrained Language Model for Scientific Text (Beltagy et al., EMNLP-IJCNLP 2019)
Modeling
Modeling
Modeling
[14] Well-read students learn better: On the importance of pre-training compact models (Turc et al., 2019)
Modeling
[15] Unified Language Model Pre-training for Natural Language Understanding and Generation (Dong et al., NeurIPS 2019)
Results
Results: KPE vs. KPG
Results: BERT for KPG – prefix-LM
Results: BERT for KPG – prefix-LM
Results: BERT for KPG – BERT2BERT
Results: BERT for KPG – BERT2BERT
Results: BERT for KPG – BERT2BERT
Summary
Thank you for listening!
References
[1] Unsupervised Keyphrase Extraction by Jointly Modeling Local and Global Context (Liang et al., EMNLP 2021)
[2] PromptRank: Unsupervised Keyphrase Extraction Using Prompt (Kong et al., ACL 2023)
[3] Learning Rich Representation of Keyphrases from Text (Kulkarni et al., Findings 2022)
[4] Representation Learning for Resource-Constrained Keyphrase Generation (Wu et al., Findings 2022)
[5] Retrieval-Augmented Multilingual Keyphrase Generation with Retriever-Generator Iterative Training (Gao et al., Findings 2022)
[6] General-to-Specific Transfer Labeling for Domain Adaptable Keyphrase Generation (Meng et al., Findings 2023)
[7] Rethinking Model Selection and Decoding for Keyphrase Generation with Pre-trained Sequence-to-Sequence Models (Wu et al., EMNLP 2023)
[8] Deep Keyphrase Generation (Meng et al., ACL 2017)
[9] KPTimes: A Large-Scale Dataset for Keyphrase Generation on News Documents (Gallina et al., INLG 2019)
[10] One Size Does Not Fit All: Generating and Evaluating Variable Number of Keyphrases (Yuan et al., ACL 2020)
[11] Exclusive Hierarchical Decoding for Deep Keyphrase Generation (Chen et al., ACL 2020)
[12] One2Set: Generating Diverse Keyphrases as a Set (Ye et al., ACL-IJCNLP 2021)
[13] SciBERT: A Pretrained Language Model for Scientific Text (Beltagy et al., EMNLP-IJCNLP 2019)
[14] Well-read students learn better: On the importance of pre-training compact models (Turc et al., 2019)
[15] Unified Language Model Pre-training for Natural Language Understanding and Generation (Dong et al., NeurIPS 2019)