File size: 1,456 Bytes
1f75c4c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
====== Perplexity statistics ======
Mean PPL(Q)                   :  24.806925 ±   0.237610
Mean PPL(base)                :  23.352232 ±   0.220841
Cor(ln(PPL(Q)), ln(PPL(base))):  99.16%
Mean ln(PPL(Q)/PPL(base))     :   0.060430 ±   0.001239
Mean PPL(Q)/PPL(base)         :   1.062293 ±   0.001316
Mean PPL(Q)-PPL(base)         :   1.454692 ±   0.034094

====== KL divergence statistics ======
Mean    KLD:   0.063828 ±   0.000254
Maximum KLD:   4.249581
99.9%   KLD:   0.920882
99.0%   KLD:   0.436885
99.0%   KLD:   0.436885
Median  KLD:   0.031026
10.0%   KLD:   0.000750
 5.0%   KLD:   0.000177
 1.0%   KLD:   0.000000
Minimum KLD:  -0.000455

====== Token probability statistics ======
Mean    Δp: -0.257 ± 0.016 %
Maximum Δp: 73.489%
99.9%   Δp: 36.078%
99.0%   Δp: 19.828%
95.0%   Δp:  8.789%
90.0%   Δp:  4.486%
75.0%   Δp:  0.496%
Median  Δp: -0.000%
25.0%   Δp: -0.687%
10.0%   Δp: -5.257%
 5.0%   Δp: -10.230%
 1.0%   Δp: -22.714%
 0.1%   Δp: -43.074%
Minimum Δp: -90.990%
RMS Δp    :  6.411 ± 0.033 %
Same top p: 88.802 ± 0.081 %

llama_perf_context_print:        load time =   87933.48 ms
llama_perf_context_print: prompt eval time = 1845252.27 ms / 304128 tokens (    6.07 ms per token,   164.82 tokens per second)
llama_perf_context_print:        eval time =       0.00 ms /     1 runs   (    0.00 ms per token,      inf tokens per second)
llama_perf_context_print:       total time = 1979075.22 ms / 304129 tokens