João Rafael commited on
Commit
8d8911a
·
1 Parent(s): 26347e8

Replace model with mpt-7b-instruct, loaded in f16 and sharded to 2GB chunks

Browse files
config.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
- "_name_or_path": "mosaicml/mpt-7b-storywriter",
3
  "architectures": [
4
  "MPTForCausalLM"
5
  ],
6
  "attn_config": {
7
  "alibi": true,
8
- "alibi_bias_max": 16,
9
  "attn_impl": "torch",
10
  "attn_pdrop": 0,
11
  "attn_type": "multihead_attention",
12
  "attn_uses_sequence_id": false,
13
- "clip_qkv": 6,
14
  "prefix_lm": false,
15
  "qk_ln": false,
16
  "softmax_scale": null
@@ -37,7 +37,7 @@
37
  "init_device": "cpu",
38
  "learned_pos_emb": true,
39
  "logit_scale": null,
40
- "max_seq_len": 65536,
41
  "model_type": "mpt",
42
  "n_heads": 32,
43
  "n_layers": 32,
@@ -46,7 +46,7 @@
46
  "resid_pdrop": 0,
47
  "tokenizer_name": "EleutherAI/gpt-neox-20b",
48
  "torch_dtype": "bfloat16",
49
- "transformers_version": "4.28.1",
50
  "use_cache": true,
51
  "verbose": 0,
52
  "vocab_size": 50432
 
1
  {
2
+ "_name_or_path": "mosaicml/mpt-7b-instruct",
3
  "architectures": [
4
  "MPTForCausalLM"
5
  ],
6
  "attn_config": {
7
  "alibi": true,
8
+ "alibi_bias_max": 8,
9
  "attn_impl": "torch",
10
  "attn_pdrop": 0,
11
  "attn_type": "multihead_attention",
12
  "attn_uses_sequence_id": false,
13
+ "clip_qkv": null,
14
  "prefix_lm": false,
15
  "qk_ln": false,
16
  "softmax_scale": null
 
37
  "init_device": "cpu",
38
  "learned_pos_emb": true,
39
  "logit_scale": null,
40
+ "max_seq_len": 2048,
41
  "model_type": "mpt",
42
  "n_heads": 32,
43
  "n_layers": 32,
 
46
  "resid_pdrop": 0,
47
  "tokenizer_name": "EleutherAI/gpt-neox-20b",
48
  "torch_dtype": "bfloat16",
49
+ "transformers_version": "4.29.2",
50
  "use_cache": true,
51
  "verbose": 0,
52
  "vocab_size": 50432
generation_config.json CHANGED
@@ -1,13 +1,5 @@
1
  {
2
- "bos_token_id":0,
3
- "_from_model_config":false,
4
- "do_sample":true,
5
- "eos_token_id":0,
6
- "max_new_tokens":512,
7
- "no_repeat_ngram_size":10,
8
- "pad_token_id":0,
9
- "penalty_alpha":0.5,
10
- "repetition_penalty":1.05,
11
- "top_k":4,
12
- "transformers_version":"4.28.1"
13
- }
 
1
  {
2
+ "_from_model_config": true,
3
+ "transformers_version": "4.29.2",
4
+ "use_cache": false
5
+ }
 
 
 
 
 
 
 
 
pytorch_model-00001-of-00007.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e06e21e5597362475c4647f78111969186a50b13e842c704f6f2aa4a4891ad16
3
- size 1889607551
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:504b9923f3ed1c8d9605c2fe470db548a5952ca8c488ee8c3b87f5ba3c8e3e88
3
+ size 1889608022
pytorch_model-00002-of-00007.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9e695959203c11763b1d19c9441279afccee5b0259099f6153d12eff1e4e11b
3
- size 1879139801
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc128efcdc9cfd2d00e88ad1686996658e0a7c53a35281bcbcfd248c09584bf4
3
+ size 1879140272
pytorch_model-00003-of-00007.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4b686f5548dd7b6e385bc4d6960edc519fc7be06070e62fd5cc8582841d109c7
3
- size 1979794567
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f13268b4cd1ae32acc89155a3f09da882951ea7b254aece34693b47986b97a68
3
+ size 1979795038
pytorch_model-00004-of-00007.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8a24a2d01c2654329961ef0479653b571be6cd3702e5398cee93af4ed7dad995
3
- size 1912694233
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36bcd56375703ea98f8e7fd571bb2188bbf1b140d39660e49bafa10b233edc6f
3
+ size 1912694704
pytorch_model-00005-of-00007.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:90425be6ec1c91e57316858f3bc4b7496863e4d938cc181b0fe7722d8f88051b
3
- size 1879131271
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d30513408b88b1a77bf1c5b88bf285751e9e163bf3866239da771c18a74d642
3
+ size 1879131806
pytorch_model-00006-of-00007.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a32c8295b8c663a339e6375cae8bf585376bdc7898d6ec3eac3aa9533ccf4bbe
3
- size 1879139801
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37653de6fb3ed72181e55e8835a610f22b74254ff9c8c3bbc2d3a801615fc3e1
3
+ size 1879140272
pytorch_model-00007-of-00007.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5fb7cbcc874e2dcba3e928017de4d60db2a2e4a1a32090a49aec68ab0a252d0b
3
- size 1879130997
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:435e5dff52b384064a2984eafe651c9aaabe466a16dda239b60cccbb4770e331
3
+ size 1879131468