This is the fourth in a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus. This model is fine-tuned on top of Mistral-Nemo-Base-2407.
Prompting
Model has been Instruct tuned with the ChatML formatting. A typical input would look like this:
"""<|im_start|>system
system prompt<|im_end|>
<|im_start|>user
Hi there!<|im_end|>
<|im_start|>assistant
Nice to meet you!<|im_end|>
<|im_start|>user
Can I ask a question?<|im_end|>
<|im_start|>assistant
"""
Credits
- Stheno dataset (filtered)
- kalomaze/Opus_Instruct_25k
- Nopm/Opus_WritingStruct
- Gryphe/Sonnet3.5-SlimOrcaDedupCleaned (A ~16k rows subset)
- kalomaze/Opus_Instruct_3k
This model has been a team effort, and the credits goes to all members of Anthracite.
Training
The training was done for 2 epochs. We used 8x NVIDIA H100 Tensor Core GPUs for the full-parameter fine-tuning of the model.
Safety
...
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 18.68 |
IFEval (0-Shot) | 37.62 |
BBH (3-Shot) | 28.79 |
MATH Lvl 5 (4-Shot) | 4.76 |
GPQA (0-shot) | 5.48 |
MuSR (0-shot) | 11.37 |
MMLU-PRO (5-shot) | 24.08 |
- Downloads last month
- 99
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
HF Inference API was unable to determine this model's library.
Model tree for mav23/magnum-v2-12b-GGUF
Base model
mistralai/Mistral-Nemo-Base-2407Evaluation results
- strict accuracy on IFEval (0-Shot)Open LLM Leaderboard37.620
- normalized accuracy on BBH (3-Shot)Open LLM Leaderboard28.790
- exact match on MATH Lvl 5 (4-Shot)Open LLM Leaderboard4.760
- acc_norm on GPQA (0-shot)Open LLM Leaderboard5.480
- acc_norm on MuSR (0-shot)Open LLM Leaderboard11.370
- accuracy on MMLU-PRO (5-shot)test set Open LLM Leaderboard24.080