mistral-doryV2-12b - EXL2 8bpw max

This is a 8bpw EXL2 quant of BeaverAI/mistral-doryV2-12b

This quant was made using exllamav2-0.1.8 with default dataset. I used a slightly modified quantization script to force use of highest bpw methods for all layers in the model (which is usually "1:8b_128g s4") to ensure max quality.

I also added a small fix in config file to set max default context at 128k as original Mistral-Nemo should have.

I tested this quant shortly in some random RPs (including ones over 8k context) and it seems to work fine.

Prompt Templates

Uses Alpaca-like format like mentioned below.

Original readme below


Dory 12b (v2)

(redone) redone instruct finetune of mistral nemo 12b's base. not (E)RP-focused, leave that to drummer.

image/gif

thanks to twisted again for the compute :3

Prompting

alpaca-like:

### System:
[Optional system prompt]

### Instruction:
[Query]

### Response:
[Response]</s>

### Instruction:
[...]

Training details

Rank 64 QDoRA, trained on the following data mix:

Downloads last month
1
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for DeusImperator/mistral-doryV2-12b_exl2_8bpw_max

Quantized
(80)
this model

Collection including DeusImperator/mistral-doryV2-12b_exl2_8bpw_max