Open LLM Leaderboard Evaluation Results
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 41.06 |
IFEval (0-Shot) | 69.06 |
BBH (3-Shot) | 56.21 |
MATH Lvl 5 (4-Shot) | 41.31 |
GPQA (0-shot) | 14.43 |
MuSR (0-shot) | 16.34 |
MMLU-PRO (5-shot) | 48.98 |
- Downloads last month
- 5
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API:
The model has no library tag.
Model tree for Quazim0t0/Ponder-14B-linear
Merge model
this model
Evaluation results
- strict accuracy on IFEval (0-Shot)Open LLM Leaderboard69.060
- normalized accuracy on BBH (3-Shot)Open LLM Leaderboard56.210
- exact match on MATH Lvl 5 (4-Shot)Open LLM Leaderboard41.310
- acc_norm on GPQA (0-shot)Open LLM Leaderboard14.430
- acc_norm on MuSR (0-shot)Open LLM Leaderboard16.340
- accuracy on MMLU-PRO (5-shot)test set Open LLM Leaderboard48.980