© dall-e
Why Does ChatGPT “Delve” So Much? Exploring the Sources of Lexical
Overrepresentation in Large Language Models
21 Jan 2025 @ COLING25
T.S. Juzek & Z. B. Ward
Joint work with �Zina Ward
Throughout the project great input by Gordon Erlebacher
Links
Background
Language changes over time
Scientific English changes over time (→ Elke Teich’s Team at Saarland University)
Examples:
Background
There have been rapid changes recently
These changes are hard to explain ‘naturally’
Background
Koppenburg, 2024; Nguyen, 2024; Shapira, 2024; Gray, 2024; Kobak et al., 2024; Liang et al., 2024; Liu and Bu, 2024; Matsui, 2024; Juzek and Ward 2025
Background
Koppenburg, 2024; Nguyen, 2024; Shapira, 2024; Gray, 2024; Kobak et al., 2024; Liang et al., 2024; Liu and Bu, 2024; Matsui, 2024; Juzek and Ward 2025
Broader impacts:
(virtually) unprecedented language change
Background
And early on, these changes were attributed to the influence of Large Language Models (LLMs) like ChatGPT
Background
However:
Background
Our work:
The procedure:
List of factors
List of factors
possible, but no strong starting points
List of factors
List of factors
language output by Llama Base (-LHF)
vs �Llama Instruct (+LHF)
→ indicator
List of factors
language output by Llama Base (-LHF)
vs �Llama Instruct (+LHF)
→ indicator
N.B.: RLHF, DPO, and LHF
illustration from Rafailov et al. 2024
Direct Preference Learning
Rafailov et al. 2024
Roof term
Learning from Human Feedback (LHF)
Because of Direct Preference Learning → Llama 3
LHF
LHF
Experiment: Emulate LHF
Analysis
Results
→virtually no chance to get conclusive experimental results
→conjecture → follow-up
Limitations
Intellectual merit
Broader impacts
Broader impacts
Broader impacts
Thank you