MasakhaPOS: Part-of-Speech Tagging for Typologically Diverse African Languages
ACL 2023
July 11, 2023
Presenter:
David Ifeoluwa Adelani (@davlanade)
UCL, United Kingdom
2
Motivation
3
Figure obtained from Joakim Nivre. Multilingual Dependency Parsing from Universal Dependencies to Sesame Street . In TSD, 2020
No large-scale POS dataset for African languages, only one Sub-Saharan African language (Wolof) has training data in UD.
Contributions
4
Languages and data size
Annotated corpus is based on the news domain
NC: Niger-Congo
5
Hausa
Kiswahili
chiShona
Setswana
isiXhosa
isiZulu
Chichewa
Kiswahili
Dholuo
Kiswahili
Luganda
Kinyarwanda
Ghomala
Wolof
Mossi
Twi
Ewe
Fon
Yoruba
Igbo
Naija
Data split roughly: 800/100/600 train/dev/test
Annotation methodology
6
Annotation Methodology
{n=1400} sentences
{n=100} sentences
Train a POS tagger (RemBERT)
Manually label
Automatic Labeling
Fix incorrect tags
https://universaldependencies.org/u/pos/
Challenges in annotating POS with UD guidelines
7
2. POS Ambiguities
Baseline models
8
Supports 2 - 8 focus languages
Supports 6 - 20 focus languages
Baseline results
Full-supervised - 800 training sentences
Multilingual pre-trained language models provides better performance than CRF
9
Baseline results
Full-supervised - 800 training sentences
Larger PLMs leads to further boost in performance
10
Baseline results
Full-supervised - 800 training sentences
African-centric PLMs that cover more African languages are more effective
11
Cross-lingual Transfer: Adapting to unseen languages
12
Transfer results: on aggregate
Geographically close source languages: Afrikaans, Arabic, English, French, Naija, Romanian and Wolof
FT-Eval: Fine-tune PLM on SOURCE, zero-shot evaluation on TARGET
Parameter-efficient fine-tuning methods:
13
Transfer results: on aggregate
Geographically close source languages: Afrikaans, Arabic, English, French, Naija, Romanian and Wolof
FT-Eval: Fine-tune PLM on a source language, perform zero-shot evaluation on the target language.
Parameter-efficient fine-tuning methods:
14
Transfer results: fine-grained analysis
How important is the source language using MAD-X adaptation?
Evaluation on: Fon, Yoruba, and isiZulu
15
Multi-source adaptation is very effective
Conclusion
16
Thank you
17
Hausa
Kiswahili
chiShona
Setswana
isiXhosa
isiZulu
Chichewa
Kiswahili
Dholuo
Kiswahili
Luganda
Kinyarwanda
Ghomala
Wolof
Mossi
Twi
Ewe
Fon
Yoruba
Igbo
Naija
MasakhaPOS
BACKUP SLIDE
18
Parameter-Efficient Fine-tuning with Adapters
19
19
STEP 1: Train language adapters on every language of interest
STEP 2: Train task adapter together with the language adapter, only modify task adapter, others parameters are frozen
Ruder 2022: NLP for African languages @Indaba
STEP 3: Zero-Shot Transfer to Target Language by replacing the source language adapter but keeping the task adapter.
Lottery Ticket Sparse Fine-tunings (LT-SFT)
20
Ansell et al. 2022. Composable Sparse Fine-Tuning for Cross-Lingual Transfer. In ACL 2022.