1
You don't need a personality test to know these models are unreliable:
Assessing the Reliability of Large Language Models on Psychometric Instruments
Bangzhao Shu*
Minje
Choi
David
Jurgens
Lechen Zhang*
Lavinia
Dunagan
Lajanugen Logeswaran
Moontae
Lee
Dallas
Card
* Equal Contribution
LLMs can be replicas of human agents
2
But do LLMs actually have consistent personas?
LLMs can be replicas of human agents
Research Questions
4
What is a Persona?
5
Do you agree that you are the life of the party?
Yes!
How Could We Measure LLMs’ Personas?
6
Statement:
<Statement> (You are the life of the party.)
Question:
Do you agree with the statement? Reply with only ‘Yes’ or ‘No’ without explaining your reasoning.
Answer:
Model-Personas: A Comprehensive Dataset for Measuring Personas
7
Model-Personas: A Comprehensive Dataset for Measuring Personas
8
Questions (693)
I am interested in people
I sympathize with others’ feelings
…
It would be okay if some people were treated differently from others
It would be okay if someone acted unfairly
…
Instruments
(39)
AIS
OCEAN
EPQ
…
MFT
MBTI
ACI
…
Persona Axes
(115)
Agreeableness
Extroversion
…
Authority
Avoid harm
Fairness
…
Which LLMs might have personas?
9
But how reliable are the LLM-generated responses?
We have instruments to measure with.
Yes
Yes
Do you agree that you are the life of the party?
No
Yes
Do you agree that you are not the life of the party?
No
Yes
Do you agree that you avoid being the life of the party?
Research Questions
11
Criteria for Examining the Reliability of LLMs
12
Do you agree?
Gracias!
GPT-2
Yes
GPT-4
Criteria for Examining the Reliability of LLMs
13
Yes
Yes
Answer?
Yes
No
Answer:
Yes
No
Answer:\n
Criteria for Examining the Reliability of LLMs
14
[Consistency] Types of content-level variation
Those consistency test are easy for people, but what about LLMs?
Results
16
Comprehensibility: Can LLMs answer with Yes or No even after spurious prompt changes?
17
Sensitivity: Do models retain their answers even after spurious prompt changes?
Sensitivity: Do models retain their answers even after spurious prompt changes?
Prompt Ending:
LLMs flip their answers after an added space
Random
Sensitivity: Do models retain their answers even after spurious prompt changes?
Prompt Ending:
Sensitivity: Do models retain their answers even after spurious prompt changes?
Prompt Ending:
The model’s architecture matters a lot!
Consistency: Do models respond consistently to content-level prompt changes?
22
Consistency: Do models respond consistently to content-level prompt changes?
23
“Yes or No”
v.s.
“No or Yes”
Consistency: Do models respond consistently to content-level prompt changes?
24
“Yes or No”
v.s.
“No or Yes”
Consistency: Do models respond consistently to content-level prompt changes?
25
Random
“Yes or No”
v.s.
“True or False”
Consistency: Do models respond consistently to content-level prompt changes?
26
Random
“Yes or No”
v.s.
“True or False”
Consistency: Do models respond consistently to content-level prompt changes?
27
Example
“should be done”
v.s.
“should not be done”
Random
Consistency: Do models respond consistently to content-level prompt changes?
28
Example
“should be done”
v.s.
“should not be done”
Random
Consistency: Do models respond consistently to content-level prompt changes?
29
Example
“should be done”
v.s.
“should be prohibited”
Random
Consistency: Do models respond consistently to content-level prompt changes?
30
Example
“should be done”
v.s.
“should be prohibited”
Random
Consistency: Do models respond consistently to content-level prompt changes?
31
LLMs struggle to provide consistent answers
32
Can we improve their consistency?
Research Questions
33
Adding personas to prompts
You are an {extroverted} person {who is outgoing and energized by interactions with other people}.
Do you agree that you are the life of the party?
Yes!
Will this happen?
Do you agree that you are the life of the party?
Maybe
34
Adding personas to prompts
35
Does adding personas improve consistency?
36
Does adding personas improve consistency?
37
Does adding personas improve consistency?
38
Summary
39
40
Bangzhao Shu*
Minje
Choi
David
Jurgens
Lechen Zhang*
Lavinia
Dunagan
Lajanugen Logeswaran
Moontae
Lee
Dallas
Card
Thanks for your listening!