File size: 1,456 Bytes
1f75c4c |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 |
====== Perplexity statistics ======
Mean PPL(Q) : 26.143038 ± 0.254443
Mean PPL(base) : 23.352232 ± 0.220841
Cor(ln(PPL(Q)), ln(PPL(base))): 98.95%
Mean ln(PPL(Q)/PPL(base)) : 0.112890 ± 0.001420
Mean PPL(Q)/PPL(base) : 1.119509 ± 0.001590
Mean PPL(Q)-PPL(base) : 2.790806 ± 0.048102
====== KL divergence statistics ======
Mean KLD: 0.082486 ± 0.000324
Maximum KLD: 8.223216
99.9% KLD: 1.163796
99.0% KLD: 0.554553
99.0% KLD: 0.554553
Median KLD: 0.040744
10.0% KLD: 0.000979
5.0% KLD: 0.000241
1.0% KLD: 0.000008
Minimum KLD: -0.000550
====== Token probability statistics ======
Mean Δp: -0.290 ± 0.018 %
Maximum Δp: 83.668%
99.9% Δp: 39.492%
99.0% Δp: 22.349%
95.0% Δp: 10.000%
90.0% Δp: 4.974%
75.0% Δp: 0.538%
Median Δp: -0.001%
25.0% Δp: -0.805%
10.0% Δp: -5.962%
5.0% Δp: -11.458%
1.0% Δp: -25.491%
0.1% Δp: -47.606%
Minimum Δp: -94.438%
RMS Δp : 7.163 ± 0.036 %
Same top p: 87.635 ± 0.085 %
llama_perf_context_print: load time = 80439.51 ms
llama_perf_context_print: prompt eval time = 1705180.19 ms / 304128 tokens ( 5.61 ms per token, 178.36 tokens per second)
llama_perf_context_print: eval time = 0.00 ms / 1 runs ( 0.00 ms per token, inf tokens per second)
llama_perf_context_print: total time = 1852832.45 ms / 304129 tokens
|