File size: 1,456 Bytes
1f75c4c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
====== Perplexity statistics ======
Mean PPL(Q)                   :  26.143038 ±   0.254443
Mean PPL(base)                :  23.352232 ±   0.220841
Cor(ln(PPL(Q)), ln(PPL(base))):  98.95%
Mean ln(PPL(Q)/PPL(base))     :   0.112890 ±   0.001420
Mean PPL(Q)/PPL(base)         :   1.119509 ±   0.001590
Mean PPL(Q)-PPL(base)         :   2.790806 ±   0.048102

====== KL divergence statistics ======
Mean    KLD:   0.082486 ±   0.000324
Maximum KLD:   8.223216
99.9%   KLD:   1.163796
99.0%   KLD:   0.554553
99.0%   KLD:   0.554553
Median  KLD:   0.040744
10.0%   KLD:   0.000979
 5.0%   KLD:   0.000241
 1.0%   KLD:   0.000008
Minimum KLD:  -0.000550

====== Token probability statistics ======
Mean    Δp: -0.290 ± 0.018 %
Maximum Δp: 83.668%
99.9%   Δp: 39.492%
99.0%   Δp: 22.349%
95.0%   Δp: 10.000%
90.0%   Δp:  4.974%
75.0%   Δp:  0.538%
Median  Δp: -0.001%
25.0%   Δp: -0.805%
10.0%   Δp: -5.962%
 5.0%   Δp: -11.458%
 1.0%   Δp: -25.491%
 0.1%   Δp: -47.606%
Minimum Δp: -94.438%
RMS Δp    :  7.163 ± 0.036 %
Same top p: 87.635 ± 0.085 %

llama_perf_context_print:        load time =   80439.51 ms
llama_perf_context_print: prompt eval time = 1705180.19 ms / 304128 tokens (    5.61 ms per token,   178.36 tokens per second)
llama_perf_context_print:        eval time =       0.00 ms /     1 runs   (    0.00 ms per token,      inf tokens per second)
llama_perf_context_print:       total time = 1852832.45 ms / 304129 tokens