Files changed (1) hide show
  1. README.md +121 -7
README.md CHANGED
@@ -1,8 +1,10 @@
1
  ---
2
  language:
3
  - en
 
4
  library_name: transformers
5
- pipeline_tag: text-generation
 
6
  datasets:
7
  - jondurbin/airoboros-2.2.1
8
  - Open-Orca/OpenOrca
@@ -10,22 +12,120 @@ datasets:
10
  - ehartford/samantha-data
11
  - CollectiveCognition/chats-data-2023-09-27
12
  - stingning/ultrachat
13
- tags:
14
- - code
15
- license: apache-2.0
16
  model-index:
17
  - name: SpeechlessCoder
18
  results:
19
  - task:
20
  type: text-generation
21
  dataset:
22
- type: openai_humaneval
23
  name: HumanEval
 
24
  metrics:
25
- - name: pass@1
26
- type: pass@1
27
  value: 0.0
 
28
  verified: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  ---
30
 
31
  <p><h1> speechless-mistral-six-in-one-7b-orth-1.0 </h1></p>
@@ -133,3 +233,17 @@ Mistral 7B is a pretrained base model and therefore does not have any moderation
133
  ## The Mistral AI Team
134
 
135
  Albert Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lélio Renard Lavaud, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed.`
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  language:
3
  - en
4
+ license: apache-2.0
5
  library_name: transformers
6
+ tags:
7
+ - code
8
  datasets:
9
  - jondurbin/airoboros-2.2.1
10
  - Open-Orca/OpenOrca
 
12
  - ehartford/samantha-data
13
  - CollectiveCognition/chats-data-2023-09-27
14
  - stingning/ultrachat
15
+ pipeline_tag: text-generation
 
 
16
  model-index:
17
  - name: SpeechlessCoder
18
  results:
19
  - task:
20
  type: text-generation
21
  dataset:
 
22
  name: HumanEval
23
+ type: openai_humaneval
24
  metrics:
25
+ - type: pass@1
 
26
  value: 0.0
27
+ name: pass@1
28
  verified: false
29
+ - task:
30
+ type: text-generation
31
+ name: Text Generation
32
+ dataset:
33
+ name: AI2 Reasoning Challenge (25-Shot)
34
+ type: ai2_arc
35
+ config: ARC-Challenge
36
+ split: test
37
+ args:
38
+ num_few_shot: 25
39
+ metrics:
40
+ - type: acc_norm
41
+ value: 22.7
42
+ name: normalized accuracy
43
+ source:
44
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=uukuguy/speechless-mistral-six-in-one-7b-orth-1.0
45
+ name: Open LLM Leaderboard
46
+ - task:
47
+ type: text-generation
48
+ name: Text Generation
49
+ dataset:
50
+ name: HellaSwag (10-Shot)
51
+ type: hellaswag
52
+ split: validation
53
+ args:
54
+ num_few_shot: 10
55
+ metrics:
56
+ - type: acc_norm
57
+ value: 25.04
58
+ name: normalized accuracy
59
+ source:
60
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=uukuguy/speechless-mistral-six-in-one-7b-orth-1.0
61
+ name: Open LLM Leaderboard
62
+ - task:
63
+ type: text-generation
64
+ name: Text Generation
65
+ dataset:
66
+ name: MMLU (5-Shot)
67
+ type: cais/mmlu
68
+ config: all
69
+ split: test
70
+ args:
71
+ num_few_shot: 5
72
+ metrics:
73
+ - type: acc
74
+ value: 23.12
75
+ name: accuracy
76
+ source:
77
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=uukuguy/speechless-mistral-six-in-one-7b-orth-1.0
78
+ name: Open LLM Leaderboard
79
+ - task:
80
+ type: text-generation
81
+ name: Text Generation
82
+ dataset:
83
+ name: TruthfulQA (0-shot)
84
+ type: truthful_qa
85
+ config: multiple_choice
86
+ split: validation
87
+ args:
88
+ num_few_shot: 0
89
+ metrics:
90
+ - type: mc2
91
+ value: 0.0
92
+ source:
93
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=uukuguy/speechless-mistral-six-in-one-7b-orth-1.0
94
+ name: Open LLM Leaderboard
95
+ - task:
96
+ type: text-generation
97
+ name: Text Generation
98
+ dataset:
99
+ name: Winogrande (5-shot)
100
+ type: winogrande
101
+ config: winogrande_xl
102
+ split: validation
103
+ args:
104
+ num_few_shot: 5
105
+ metrics:
106
+ - type: acc
107
+ value: 49.57
108
+ name: accuracy
109
+ source:
110
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=uukuguy/speechless-mistral-six-in-one-7b-orth-1.0
111
+ name: Open LLM Leaderboard
112
+ - task:
113
+ type: text-generation
114
+ name: Text Generation
115
+ dataset:
116
+ name: GSM8k (5-shot)
117
+ type: gsm8k
118
+ config: main
119
+ split: test
120
+ args:
121
+ num_few_shot: 5
122
+ metrics:
123
+ - type: acc
124
+ value: 0.0
125
+ name: accuracy
126
+ source:
127
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=uukuguy/speechless-mistral-six-in-one-7b-orth-1.0
128
+ name: Open LLM Leaderboard
129
  ---
130
 
131
  <p><h1> speechless-mistral-six-in-one-7b-orth-1.0 </h1></p>
 
233
  ## The Mistral AI Team
234
 
235
  Albert Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lélio Renard Lavaud, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed.`
236
+
237
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
238
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_uukuguy__speechless-mistral-six-in-one-7b-orth-1.0)
239
+
240
+ | Metric |Value|
241
+ |---------------------------------|----:|
242
+ |Avg. |20.07|
243
+ |AI2 Reasoning Challenge (25-Shot)|22.70|
244
+ |HellaSwag (10-Shot) |25.04|
245
+ |MMLU (5-Shot) |23.12|
246
+ |TruthfulQA (0-shot) | 0.00|
247
+ |Winogrande (5-shot) |49.57|
248
+ |GSM8k (5-shot) | 0.00|
249
+