ABCDEFGHIJKLMNOPQRSTUVWXYZ
1
Shannon game results -- CS 585, 9/15/16
2
this sheet:
brenocon.com/2016ex1
3
materials:
http://people.cs.umass.edu/~brenocon/inlp2016/ex1_lm.pdf
4
Place your score in the column corresponding to your worksheet version.
5
score = log10 prob of revealed word
NOTE: this isn't really perplexity since typically we take 1 model and evaluate on N tokens.
6
instead we are taking ~30 models (number of people per version) and evaluating on the same data point over and over.
7
Version 1Version 2Version 3
we're just doing this to illustrate how the arithmetic works.
8
(the _)(load it into the _)
(two sentences of context)
DATA CHECK: positive logprobs are NOT POSSIBLE. We had to delete a bunch people put in.
9
-0.22-0.7-0.3avg logprobneg avg logprob
exp(neg avg logprob) (a.k.a. PERPLEXITY)
10
-0.523-0.2218-0.2Version 1-0.85979163040.85979163047.240884672
11
-0.7-3-0.00436Version 2-0.59473523870.59473523873.933102265
12
-0.602059991-0.5-0.15Version 3-0.42392786430.42392786432.654164672
13
-0.3979400087-0.3979400087-0.22
14
-2-0.46
is the worst theoretical ppl
15
-0.69897-0.30103-0.12
(if you predicted 1/6 for each word, for V=6)
16
-0.5
17
-0.69-0.5
NOTE: in class we did log base 10, but in the spreadsheet during class I typed exp() to calc PPL.
18
-2.522878745-0.3
but it should have been 10^x since the log was base 10.
19
-0.7-0.3010299957
ppl = exp(-log_e prob) = 2^(-log_2 prob) = 10^(-log_10 prob)
20
-0.69897-0.35-0.3979400087
21
-0.5-1.5-0.31
I corrected it after class. now Version 1 performance was worse than the theoretical baseline!
22
-0.5228787453-0.7425-0.398
why is this possible?
23
-0.5228-0.5333-0.52
24
-0.3979400087-0.1249
25
-0.301-0.523
26
-0.69897-0.22
27
-2.5-0.3979400087-0.5228
28
-0.4-1-0.05
29
30
-0.301
31
-1
32
33
-10.096
34
-0.456-0.0457575
35
-0.2596373105
36
-0.5228-0.3
37
-2-2.99
38
39
-1.2
40
-0.0087739243
41
-0.398-0.301
42
-0.398
43
-0.05
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100