====== Perplexity statistics ====== Mean PPL(Q) : 24.415348 ± 0.234385 Mean PPL(base) : 23.352232 ± 0.220841 Cor(ln(PPL(Q)), ln(PPL(base))): 99.79% Mean ln(PPL(Q)/PPL(base)) : 0.044519 ± 0.000634 Mean PPL(Q)/PPL(base) : 1.045525 ± 0.000663 Mean PPL(Q)-PPL(base) : 1.063116 ± 0.020031 ====== KL divergence statistics ====== Mean KLD: 0.015503 ± 0.000075 Maximum KLD: 3.272221 99.9% KLD: 0.267841 99.0% KLD: 0.107995 99.0% KLD: 0.107995 Median KLD: 0.007291 10.0% KLD: 0.000184 5.0% KLD: 0.000033 1.0% KLD: -0.000041 Minimum KLD: -0.000471 ====== Token probability statistics ====== Mean Δp: -0.008 ± 0.008 % Maximum Δp: 80.786% 99.9% Δp: 18.971% 99.0% Δp: 10.037% 95.0% Δp: 4.585% 90.0% Δp: 2.431% 75.0% Δp: 0.317% Median Δp: -0.000% 25.0% Δp: -0.289% 10.0% Δp: -2.327% 5.0% Δp: -4.593% 1.0% Δp: -10.622% 0.1% Δp: -22.631% Minimum Δp: -95.612% RMS Δp : 3.201 ± 0.022 % Same top p: 94.448 ± 0.059 % llama_perf_context_print: load time = 134148.25 ms llama_perf_context_print: prompt eval time = 2502976.05 ms / 304128 tokens ( 8.23 ms per token, 121.51 tokens per second) llama_perf_context_print: eval time = 0.00 ms / 1 runs ( 0.00 ms per token, inf tokens per second) llama_perf_context_print: total time = 3009280.06 ms / 304129 tokens