João Rafael
commited on
Commit
·
8d8911a
1
Parent(s):
26347e8
Replace model with mpt-7b-instruct, loaded in f16 and sharded to 2GB chunks
Browse files- config.json +5 -5
- generation_config.json +4 -12
- pytorch_model-00001-of-00007.bin +2 -2
- pytorch_model-00002-of-00007.bin +2 -2
- pytorch_model-00003-of-00007.bin +2 -2
- pytorch_model-00004-of-00007.bin +2 -2
- pytorch_model-00005-of-00007.bin +2 -2
- pytorch_model-00006-of-00007.bin +2 -2
- pytorch_model-00007-of-00007.bin +2 -2
config.json
CHANGED
@@ -1,16 +1,16 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "mosaicml/mpt-7b-
|
3 |
"architectures": [
|
4 |
"MPTForCausalLM"
|
5 |
],
|
6 |
"attn_config": {
|
7 |
"alibi": true,
|
8 |
-
"alibi_bias_max":
|
9 |
"attn_impl": "torch",
|
10 |
"attn_pdrop": 0,
|
11 |
"attn_type": "multihead_attention",
|
12 |
"attn_uses_sequence_id": false,
|
13 |
-
"clip_qkv":
|
14 |
"prefix_lm": false,
|
15 |
"qk_ln": false,
|
16 |
"softmax_scale": null
|
@@ -37,7 +37,7 @@
|
|
37 |
"init_device": "cpu",
|
38 |
"learned_pos_emb": true,
|
39 |
"logit_scale": null,
|
40 |
-
"max_seq_len":
|
41 |
"model_type": "mpt",
|
42 |
"n_heads": 32,
|
43 |
"n_layers": 32,
|
@@ -46,7 +46,7 @@
|
|
46 |
"resid_pdrop": 0,
|
47 |
"tokenizer_name": "EleutherAI/gpt-neox-20b",
|
48 |
"torch_dtype": "bfloat16",
|
49 |
-
"transformers_version": "4.
|
50 |
"use_cache": true,
|
51 |
"verbose": 0,
|
52 |
"vocab_size": 50432
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "mosaicml/mpt-7b-instruct",
|
3 |
"architectures": [
|
4 |
"MPTForCausalLM"
|
5 |
],
|
6 |
"attn_config": {
|
7 |
"alibi": true,
|
8 |
+
"alibi_bias_max": 8,
|
9 |
"attn_impl": "torch",
|
10 |
"attn_pdrop": 0,
|
11 |
"attn_type": "multihead_attention",
|
12 |
"attn_uses_sequence_id": false,
|
13 |
+
"clip_qkv": null,
|
14 |
"prefix_lm": false,
|
15 |
"qk_ln": false,
|
16 |
"softmax_scale": null
|
|
|
37 |
"init_device": "cpu",
|
38 |
"learned_pos_emb": true,
|
39 |
"logit_scale": null,
|
40 |
+
"max_seq_len": 2048,
|
41 |
"model_type": "mpt",
|
42 |
"n_heads": 32,
|
43 |
"n_layers": 32,
|
|
|
46 |
"resid_pdrop": 0,
|
47 |
"tokenizer_name": "EleutherAI/gpt-neox-20b",
|
48 |
"torch_dtype": "bfloat16",
|
49 |
+
"transformers_version": "4.29.2",
|
50 |
"use_cache": true,
|
51 |
"verbose": 0,
|
52 |
"vocab_size": 50432
|
generation_config.json
CHANGED
@@ -1,13 +1,5 @@
|
|
1 |
{
|
2 |
-
"
|
3 |
-
"
|
4 |
-
"
|
5 |
-
|
6 |
-
"max_new_tokens":512,
|
7 |
-
"no_repeat_ngram_size":10,
|
8 |
-
"pad_token_id":0,
|
9 |
-
"penalty_alpha":0.5,
|
10 |
-
"repetition_penalty":1.05,
|
11 |
-
"top_k":4,
|
12 |
-
"transformers_version":"4.28.1"
|
13 |
-
}
|
|
|
1 |
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"transformers_version": "4.29.2",
|
4 |
+
"use_cache": false
|
5 |
+
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
pytorch_model-00001-of-00007.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:504b9923f3ed1c8d9605c2fe470db548a5952ca8c488ee8c3b87f5ba3c8e3e88
|
3 |
+
size 1889608022
|
pytorch_model-00002-of-00007.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dc128efcdc9cfd2d00e88ad1686996658e0a7c53a35281bcbcfd248c09584bf4
|
3 |
+
size 1879140272
|
pytorch_model-00003-of-00007.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f13268b4cd1ae32acc89155a3f09da882951ea7b254aece34693b47986b97a68
|
3 |
+
size 1979795038
|
pytorch_model-00004-of-00007.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:36bcd56375703ea98f8e7fd571bb2188bbf1b140d39660e49bafa10b233edc6f
|
3 |
+
size 1912694704
|
pytorch_model-00005-of-00007.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d30513408b88b1a77bf1c5b88bf285751e9e163bf3866239da771c18a74d642
|
3 |
+
size 1879131806
|
pytorch_model-00006-of-00007.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:37653de6fb3ed72181e55e8835a610f22b74254ff9c8c3bbc2d3a801615fc3e1
|
3 |
+
size 1879140272
|
pytorch_model-00007-of-00007.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:435e5dff52b384064a2984eafe651c9aaabe466a16dda239b60cccbb4770e331
|
3 |
+
size 1879131468
|