aashish1904 commited on
Commit
bb074a7
·
verified ·
1 Parent(s): e791bff

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +89 -0
README.md ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ base_model: mistralai/Mistral-Nemo-Base-2407
5
+ license: cc-by-nc-4.0
6
+ tags:
7
+ - general-purpose
8
+ - roleplay
9
+ - storywriting
10
+ - merge
11
+ - finetune
12
+ library_name: transformers
13
+
14
+ ---
15
+
16
+ ![](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)
17
+
18
+ # QuantFactory/Chronos-Gold-12B-1.0-GGUF
19
+ This is quantized version of [elinas/Chronos-Gold-12B-1.0](https://huggingface.co/elinas/Chronos-Gold-12B-1.0) created using llama.cpp
20
+
21
+ # Original Model Card
22
+
23
+
24
+ # Chronos Gold 12B-1.0
25
+
26
+ ![image/webp](https://cdn-uploads.huggingface.co/production/uploads/630417380907b9a115c6aa9f/3hc8zt8fzKdO3qHK1p1mW.webp)
27
+
28
+ Chronos Gold 12B 1.0 is a very unique model that applies to domain areas such as
29
+ geneal chatbot functionatliy, *roleplay*, and storywriting. The model has been observed to write up to 2250 tokens in a single sequence. The model was trained at a
30
+ sequence length of 16384 (16k) and will still retain the *apparent* 128k context length from Mistral-Nemo.
31
+
32
+ The base model is `mistralai/Mistral-Nemo-Base-2407` which was heavily modified to produce a more coherent model, comparable to much larger models.
33
+
34
+ **Chronos Gold 12B-1.0** re-creates the uniqueness of the original Chronos with significiantly enhanced prompt adherence (following), coherence, a modern dataset, as well as supporting a majority of "character card" formats in applications like SillyTavern.
35
+
36
+ It went through an iterative and objective merge process as my previous models and was further finetuned on a dataset curated for it.
37
+
38
+ The specifics of the model will not be disclosed at the time due to dataset ownership.
39
+
40
+ ## Instruct Template
41
+
42
+ This model uses `ChatML` - below is an example. It is a preset in many frontends.
43
+
44
+ ```
45
+ <|im_start|>system
46
+ A system prompt describing how you'd like your bot to act.<|im_end|>
47
+ <|im_start|>user
48
+ Hello there!<|im_end|>
49
+ <|im_start|>assistant
50
+ I can assist you or we can disucss other things?<|im_end|>
51
+ <|im_start|>user
52
+ I was wondering how transformers work?<|im_end|>
53
+ <|im_start|>assistant
54
+ ```
55
+
56
+ ## Quantization
57
+
58
+ #### LlamaCPP
59
+ [GGUFs from @bartowski](https://huggingface.co/bartowski/Chronos-Gold-12B-1.0-GGUF)
60
+
61
+ #### Exllama2
62
+ [4.5bpw by @Pyroserenus](https://huggingface.co/Pyroserenus/Chronos-Gold-12B-1.0-4.5bpw-h6-exl2)
63
+
64
+ [5.5bpw by @Pyroserenus](https://huggingface.co/Pyroserenus/Chronos-Gold-12B-1.0-5.5bpw-h6-exl2)
65
+
66
+ [6.5bpw by @Pyroserenus](https://huggingface.co/Pyroserenus/Chronos-Gold-12B-1.0-6.5bpw-h6-exl2)
67
+
68
+ [8.0bpw by @Pyroserenus](https://huggingface.co/Pyroserenus/Chronos-Gold-12B-1.0-8.0bpw-h8-exl2)
69
+
70
+ #### FP8
71
+ [FP8 Quant by @Pyroserenus](https://huggingface.co/Pyroserenus/Chronos-Gold-12B-1.0-FP8)
72
+
73
+ ## Sampling Settings
74
+ Nemo is a bit sensitive to high temperatures, so I use lower. Here are my settings:
75
+ ```
76
+ Temp - 0.7 (0.9 max)
77
+ Presence Penalty - 1.0
78
+ Repetition Penalty range - 2800
79
+ Min P - 0.10
80
+ ```
81
+
82
+ ## Additional Details
83
+ This model was created by **elinas** on discord.
84
+
85
+ This is one of multiple models to come out in the series by size and model architecture, so look forward to it!
86
+
87
+ Please be mindful of the license. This is strictly non-commercial, but free to use at your own leisure personally. Outputs generated by the model are not reflective of my views.
88
+
89
+ Contact me on Discord for inquiries.