aashish1904
commited on
Upload README.md with huggingface_hub
Browse files
README.md
ADDED
@@ -0,0 +1,89 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
---
|
3 |
+
|
4 |
+
base_model: mistralai/Mistral-Nemo-Base-2407
|
5 |
+
license: cc-by-nc-4.0
|
6 |
+
tags:
|
7 |
+
- general-purpose
|
8 |
+
- roleplay
|
9 |
+
- storywriting
|
10 |
+
- merge
|
11 |
+
- finetune
|
12 |
+
library_name: transformers
|
13 |
+
|
14 |
+
---
|
15 |
+
|
16 |
+
![](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)
|
17 |
+
|
18 |
+
# QuantFactory/Chronos-Gold-12B-1.0-GGUF
|
19 |
+
This is quantized version of [elinas/Chronos-Gold-12B-1.0](https://huggingface.co/elinas/Chronos-Gold-12B-1.0) created using llama.cpp
|
20 |
+
|
21 |
+
# Original Model Card
|
22 |
+
|
23 |
+
|
24 |
+
# Chronos Gold 12B-1.0
|
25 |
+
|
26 |
+
![image/webp](https://cdn-uploads.huggingface.co/production/uploads/630417380907b9a115c6aa9f/3hc8zt8fzKdO3qHK1p1mW.webp)
|
27 |
+
|
28 |
+
Chronos Gold 12B 1.0 is a very unique model that applies to domain areas such as
|
29 |
+
geneal chatbot functionatliy, *roleplay*, and storywriting. The model has been observed to write up to 2250 tokens in a single sequence. The model was trained at a
|
30 |
+
sequence length of 16384 (16k) and will still retain the *apparent* 128k context length from Mistral-Nemo.
|
31 |
+
|
32 |
+
The base model is `mistralai/Mistral-Nemo-Base-2407` which was heavily modified to produce a more coherent model, comparable to much larger models.
|
33 |
+
|
34 |
+
**Chronos Gold 12B-1.0** re-creates the uniqueness of the original Chronos with significiantly enhanced prompt adherence (following), coherence, a modern dataset, as well as supporting a majority of "character card" formats in applications like SillyTavern.
|
35 |
+
|
36 |
+
It went through an iterative and objective merge process as my previous models and was further finetuned on a dataset curated for it.
|
37 |
+
|
38 |
+
The specifics of the model will not be disclosed at the time due to dataset ownership.
|
39 |
+
|
40 |
+
## Instruct Template
|
41 |
+
|
42 |
+
This model uses `ChatML` - below is an example. It is a preset in many frontends.
|
43 |
+
|
44 |
+
```
|
45 |
+
<|im_start|>system
|
46 |
+
A system prompt describing how you'd like your bot to act.<|im_end|>
|
47 |
+
<|im_start|>user
|
48 |
+
Hello there!<|im_end|>
|
49 |
+
<|im_start|>assistant
|
50 |
+
I can assist you or we can disucss other things?<|im_end|>
|
51 |
+
<|im_start|>user
|
52 |
+
I was wondering how transformers work?<|im_end|>
|
53 |
+
<|im_start|>assistant
|
54 |
+
```
|
55 |
+
|
56 |
+
## Quantization
|
57 |
+
|
58 |
+
#### LlamaCPP
|
59 |
+
[GGUFs from @bartowski](https://huggingface.co/bartowski/Chronos-Gold-12B-1.0-GGUF)
|
60 |
+
|
61 |
+
#### Exllama2
|
62 |
+
[4.5bpw by @Pyroserenus](https://huggingface.co/Pyroserenus/Chronos-Gold-12B-1.0-4.5bpw-h6-exl2)
|
63 |
+
|
64 |
+
[5.5bpw by @Pyroserenus](https://huggingface.co/Pyroserenus/Chronos-Gold-12B-1.0-5.5bpw-h6-exl2)
|
65 |
+
|
66 |
+
[6.5bpw by @Pyroserenus](https://huggingface.co/Pyroserenus/Chronos-Gold-12B-1.0-6.5bpw-h6-exl2)
|
67 |
+
|
68 |
+
[8.0bpw by @Pyroserenus](https://huggingface.co/Pyroserenus/Chronos-Gold-12B-1.0-8.0bpw-h8-exl2)
|
69 |
+
|
70 |
+
#### FP8
|
71 |
+
[FP8 Quant by @Pyroserenus](https://huggingface.co/Pyroserenus/Chronos-Gold-12B-1.0-FP8)
|
72 |
+
|
73 |
+
## Sampling Settings
|
74 |
+
Nemo is a bit sensitive to high temperatures, so I use lower. Here are my settings:
|
75 |
+
```
|
76 |
+
Temp - 0.7 (0.9 max)
|
77 |
+
Presence Penalty - 1.0
|
78 |
+
Repetition Penalty range - 2800
|
79 |
+
Min P - 0.10
|
80 |
+
```
|
81 |
+
|
82 |
+
## Additional Details
|
83 |
+
This model was created by **elinas** on discord.
|
84 |
+
|
85 |
+
This is one of multiple models to come out in the series by size and model architecture, so look forward to it!
|
86 |
+
|
87 |
+
Please be mindful of the license. This is strictly non-commercial, but free to use at your own leisure personally. Outputs generated by the model are not reflective of my views.
|
88 |
+
|
89 |
+
Contact me on Discord for inquiries.
|