GGUF importance matrix (imatrix) quants for https://huggingface.co/codefuse-ai/CodeFuse-DeepSeek-33B
- The importance matrix was trained for ~50K tokens (105 batches of 512 tokens) using a general purpose imatrix calibration dataset.
- The imatrix is being used on the K-quants as well.
Layers | Context | Template |
---|---|---|
62 |
16384 |
<s>system |
- Downloads last month
- 10
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the HF Inference API does not support gguf models with pipeline type text-generation