Upload Phi4MMForCausalLM
Browse filesUpdated weights (4 epochs)
- README.md +16 -16
- config.json +1 -1
- model-00001-of-00003.safetensors +2 -2
- model-00002-of-00003.safetensors +2 -2
- model-00003-of-00003.safetensors +2 -2
- model.safetensors.index.json +63 -63
README.md
CHANGED
@@ -8,45 +8,45 @@ metrics:
|
|
8 |
- cer
|
9 |
base_model:
|
10 |
- microsoft/Phi-4-multimodal-instruct
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
model-index:
|
12 |
- name: Phi-4-mm-inst-zeroth-kor
|
13 |
results:
|
14 |
- task:
|
15 |
type: speech-to-text-translation
|
16 |
dataset:
|
17 |
-
type: seastar105/fleurs_ko_en_test
|
18 |
name: fleurs (ko-en test intersection)
|
|
|
19 |
metrics:
|
20 |
- type: bleu
|
21 |
-
name: ko2en
|
22 |
value: 7.03
|
|
|
23 |
- type: bleu
|
24 |
-
name: ko2en-cot
|
25 |
value: 7.04
|
|
|
26 |
- type: bleu
|
|
|
27 |
name: en2ko (ko-mecab)
|
28 |
-
value: 12.50
|
29 |
- type: bleu
|
30 |
-
name: en2ko-cot (ko-mecab)
|
31 |
value: 9.54
|
|
|
32 |
- task:
|
33 |
type: automatic-speech-recognition
|
34 |
dataset:
|
35 |
-
type: kresnik/zeroth_korean
|
36 |
name: zeroth_korean test
|
|
|
37 |
metrics:
|
38 |
- type: cer
|
39 |
-
name: test CER
|
40 |
value: 7.02
|
41 |
-
|
42 |
-
- ko
|
43 |
-
license: mit
|
44 |
-
tags:
|
45 |
-
- korean
|
46 |
-
- stt
|
47 |
-
- custom_code
|
48 |
-
- phi
|
49 |
-
- phi-4-multimodal
|
50 |
---
|
51 |
|
52 |
# Phi-4-multimodal-finetune-ko-speech
|
|
|
8 |
- cer
|
9 |
base_model:
|
10 |
- microsoft/Phi-4-multimodal-instruct
|
11 |
+
language:
|
12 |
+
- ko
|
13 |
+
license: mit
|
14 |
+
tags:
|
15 |
+
- korean
|
16 |
+
- stt
|
17 |
+
- custom_code
|
18 |
+
- phi
|
19 |
+
- phi-4-multimodal
|
20 |
model-index:
|
21 |
- name: Phi-4-mm-inst-zeroth-kor
|
22 |
results:
|
23 |
- task:
|
24 |
type: speech-to-text-translation
|
25 |
dataset:
|
|
|
26 |
name: fleurs (ko-en test intersection)
|
27 |
+
type: seastar105/fleurs_ko_en_test
|
28 |
metrics:
|
29 |
- type: bleu
|
|
|
30 |
value: 7.03
|
31 |
+
name: ko2en
|
32 |
- type: bleu
|
|
|
33 |
value: 7.04
|
34 |
+
name: ko2en-cot
|
35 |
- type: bleu
|
36 |
+
value: 12.5
|
37 |
name: en2ko (ko-mecab)
|
|
|
38 |
- type: bleu
|
|
|
39 |
value: 9.54
|
40 |
+
name: en2ko-cot (ko-mecab)
|
41 |
- task:
|
42 |
type: automatic-speech-recognition
|
43 |
dataset:
|
|
|
44 |
name: zeroth_korean test
|
45 |
+
type: kresnik/zeroth_korean
|
46 |
metrics:
|
47 |
- type: cer
|
|
|
48 |
value: 7.02
|
49 |
+
name: test CER
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
50 |
---
|
51 |
|
52 |
# Phi-4-multimodal-finetune-ko-speech
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "./
|
3 |
"architectures": [
|
4 |
"Phi4MMForCausalLM"
|
5 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "./Phi-4-multimodal-finetune-ko-speech",
|
3 |
"architectures": [
|
4 |
"Phi4MMForCausalLM"
|
5 |
],
|
model-00001-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0a1e57956624d5a96c5dcaab437261353efbcfaf5eff49d58d92329311020802
|
3 |
+
size 4997504848
|
model-00002-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a7ae17c45b5ab41c7659784e8cb2d6a79e6f886783b3c2af0297448a1f240c1c
|
3 |
+
size 4952333128
|
model-00003-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7277ccd89f1b88a99dd7ce37095510f2036e18e6289b732164704ee0d2934b92
|
3 |
+
size 1199389232
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"model.embed_tokens.weight": "model-00001-of-00003.safetensors",
|
@@ -1741,50 +1741,50 @@
|
|
1741 |
"model.layers.24.self_attn.qkv_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1742 |
"model.layers.24.self_attn.qkv_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1743 |
"model.layers.24.self_attn.qkv_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1744 |
-
"model.layers.25.input_layernorm.weight": "model-
|
1745 |
-
"model.layers.25.mlp.down_proj.base_layer.weight": "model-
|
1746 |
-
"model.layers.25.mlp.down_proj.lora_A.speech.weight": "model-
|
1747 |
-
"model.layers.25.mlp.down_proj.lora_A.vision.weight": "model-
|
1748 |
-
"model.layers.25.mlp.down_proj.lora_B.speech.weight": "model-
|
1749 |
-
"model.layers.25.mlp.down_proj.lora_B.vision.weight": "model-
|
1750 |
-
"model.layers.25.mlp.gate_up_proj.base_layer.weight": "model-
|
1751 |
-
"model.layers.25.mlp.gate_up_proj.lora_A.speech.weight": "model-
|
1752 |
-
"model.layers.25.mlp.gate_up_proj.lora_A.vision.weight": "model-
|
1753 |
-
"model.layers.25.mlp.gate_up_proj.lora_B.speech.weight": "model-
|
1754 |
-
"model.layers.25.mlp.gate_up_proj.lora_B.vision.weight": "model-
|
1755 |
-
"model.layers.25.post_attention_layernorm.weight": "model-
|
1756 |
"model.layers.25.self_attn.o_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1757 |
"model.layers.25.self_attn.o_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1758 |
"model.layers.25.self_attn.o_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1759 |
"model.layers.25.self_attn.o_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1760 |
"model.layers.25.self_attn.o_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1761 |
-
"model.layers.25.self_attn.qkv_proj.base_layer.weight": "model-
|
1762 |
-
"model.layers.25.self_attn.qkv_proj.lora_A.speech.weight": "model-
|
1763 |
-
"model.layers.25.self_attn.qkv_proj.lora_A.vision.weight": "model-
|
1764 |
-
"model.layers.25.self_attn.qkv_proj.lora_B.speech.weight": "model-
|
1765 |
-
"model.layers.25.self_attn.qkv_proj.lora_B.vision.weight": "model-
|
1766 |
-
"model.layers.26.input_layernorm.weight": "model-
|
1767 |
-
"model.layers.26.mlp.down_proj.base_layer.weight": "model-
|
1768 |
-
"model.layers.26.mlp.down_proj.lora_A.speech.weight": "model-
|
1769 |
-
"model.layers.26.mlp.down_proj.lora_A.vision.weight": "model-
|
1770 |
-
"model.layers.26.mlp.down_proj.lora_B.speech.weight": "model-
|
1771 |
-
"model.layers.26.mlp.down_proj.lora_B.vision.weight": "model-
|
1772 |
-
"model.layers.26.mlp.gate_up_proj.base_layer.weight": "model-
|
1773 |
-
"model.layers.26.mlp.gate_up_proj.lora_A.speech.weight": "model-
|
1774 |
-
"model.layers.26.mlp.gate_up_proj.lora_A.vision.weight": "model-
|
1775 |
-
"model.layers.26.mlp.gate_up_proj.lora_B.speech.weight": "model-
|
1776 |
-
"model.layers.26.mlp.gate_up_proj.lora_B.vision.weight": "model-
|
1777 |
-
"model.layers.26.post_attention_layernorm.weight": "model-
|
1778 |
-
"model.layers.26.self_attn.o_proj.base_layer.weight": "model-
|
1779 |
-
"model.layers.26.self_attn.o_proj.lora_A.speech.weight": "model-
|
1780 |
-
"model.layers.26.self_attn.o_proj.lora_A.vision.weight": "model-
|
1781 |
-
"model.layers.26.self_attn.o_proj.lora_B.speech.weight": "model-
|
1782 |
-
"model.layers.26.self_attn.o_proj.lora_B.vision.weight": "model-
|
1783 |
-
"model.layers.26.self_attn.qkv_proj.base_layer.weight": "model-
|
1784 |
-
"model.layers.26.self_attn.qkv_proj.lora_A.speech.weight": "model-
|
1785 |
-
"model.layers.26.self_attn.qkv_proj.lora_A.vision.weight": "model-
|
1786 |
-
"model.layers.26.self_attn.qkv_proj.lora_B.speech.weight": "model-
|
1787 |
-
"model.layers.26.self_attn.qkv_proj.lora_B.vision.weight": "model-
|
1788 |
"model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1789 |
"model.layers.27.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1790 |
"model.layers.27.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
@@ -1797,16 +1797,16 @@
|
|
1797 |
"model.layers.27.mlp.gate_up_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1798 |
"model.layers.27.mlp.gate_up_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
|
1799 |
"model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
1800 |
-
"model.layers.27.self_attn.o_proj.base_layer.weight": "model-
|
1801 |
-
"model.layers.27.self_attn.o_proj.lora_A.speech.weight": "model-
|
1802 |
-
"model.layers.27.self_attn.o_proj.lora_A.vision.weight": "model-
|
1803 |
-
"model.layers.27.self_attn.o_proj.lora_B.speech.weight": "model-
|
1804 |
-
"model.layers.27.self_attn.o_proj.lora_B.vision.weight": "model-
|
1805 |
-
"model.layers.27.self_attn.qkv_proj.base_layer.weight": "model-
|
1806 |
-
"model.layers.27.self_attn.qkv_proj.lora_A.speech.weight": "model-
|
1807 |
-
"model.layers.27.self_attn.qkv_proj.lora_A.vision.weight": "model-
|
1808 |
-
"model.layers.27.self_attn.qkv_proj.lora_B.speech.weight": "model-
|
1809 |
-
"model.layers.27.self_attn.qkv_proj.lora_B.vision.weight": "model-
|
1810 |
"model.layers.28.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1811 |
"model.layers.28.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1812 |
"model.layers.28.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
@@ -1989,22 +1989,22 @@
|
|
1989 |
"model.layers.7.mlp.down_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1990 |
"model.layers.7.mlp.down_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1991 |
"model.layers.7.mlp.down_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1992 |
-
"model.layers.7.mlp.gate_up_proj.base_layer.weight": "model-
|
1993 |
-
"model.layers.7.mlp.gate_up_proj.lora_A.speech.weight": "model-
|
1994 |
-
"model.layers.7.mlp.gate_up_proj.lora_A.vision.weight": "model-
|
1995 |
"model.layers.7.mlp.gate_up_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1996 |
-
"model.layers.7.mlp.gate_up_proj.lora_B.vision.weight": "model-
|
1997 |
"model.layers.7.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
1998 |
"model.layers.7.self_attn.o_proj.base_layer.weight": "model-00001-of-00003.safetensors",
|
1999 |
-
"model.layers.7.self_attn.o_proj.lora_A.speech.weight": "model-
|
2000 |
-
"model.layers.7.self_attn.o_proj.lora_A.vision.weight": "model-
|
2001 |
-
"model.layers.7.self_attn.o_proj.lora_B.speech.weight": "model-
|
2002 |
-
"model.layers.7.self_attn.o_proj.lora_B.vision.weight": "model-
|
2003 |
-
"model.layers.7.self_attn.qkv_proj.base_layer.weight": "model-
|
2004 |
-
"model.layers.7.self_attn.qkv_proj.lora_A.speech.weight": "model-
|
2005 |
-
"model.layers.7.self_attn.qkv_proj.lora_A.vision.weight": "model-
|
2006 |
-
"model.layers.7.self_attn.qkv_proj.lora_B.speech.weight": "model-
|
2007 |
-
"model.layers.7.self_attn.qkv_proj.lora_B.vision.weight": "model-
|
2008 |
"model.layers.8.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
2009 |
"model.layers.8.mlp.down_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
2010 |
"model.layers.8.mlp.down_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 11148920768
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"model.embed_tokens.weight": "model-00001-of-00003.safetensors",
|
|
|
1741 |
"model.layers.24.self_attn.qkv_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1742 |
"model.layers.24.self_attn.qkv_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1743 |
"model.layers.24.self_attn.qkv_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1744 |
+
"model.layers.25.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
1745 |
+
"model.layers.25.mlp.down_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1746 |
+
"model.layers.25.mlp.down_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1747 |
+
"model.layers.25.mlp.down_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1748 |
+
"model.layers.25.mlp.down_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1749 |
+
"model.layers.25.mlp.down_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1750 |
+
"model.layers.25.mlp.gate_up_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1751 |
+
"model.layers.25.mlp.gate_up_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1752 |
+
"model.layers.25.mlp.gate_up_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1753 |
+
"model.layers.25.mlp.gate_up_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1754 |
+
"model.layers.25.mlp.gate_up_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1755 |
+
"model.layers.25.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
1756 |
"model.layers.25.self_attn.o_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1757 |
"model.layers.25.self_attn.o_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1758 |
"model.layers.25.self_attn.o_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1759 |
"model.layers.25.self_attn.o_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1760 |
"model.layers.25.self_attn.o_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1761 |
+
"model.layers.25.self_attn.qkv_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1762 |
+
"model.layers.25.self_attn.qkv_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1763 |
+
"model.layers.25.self_attn.qkv_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1764 |
+
"model.layers.25.self_attn.qkv_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1765 |
+
"model.layers.25.self_attn.qkv_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1766 |
+
"model.layers.26.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
1767 |
+
"model.layers.26.mlp.down_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1768 |
+
"model.layers.26.mlp.down_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1769 |
+
"model.layers.26.mlp.down_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1770 |
+
"model.layers.26.mlp.down_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1771 |
+
"model.layers.26.mlp.down_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1772 |
+
"model.layers.26.mlp.gate_up_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1773 |
+
"model.layers.26.mlp.gate_up_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1774 |
+
"model.layers.26.mlp.gate_up_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1775 |
+
"model.layers.26.mlp.gate_up_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1776 |
+
"model.layers.26.mlp.gate_up_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1777 |
+
"model.layers.26.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
1778 |
+
"model.layers.26.self_attn.o_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1779 |
+
"model.layers.26.self_attn.o_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1780 |
+
"model.layers.26.self_attn.o_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1781 |
+
"model.layers.26.self_attn.o_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1782 |
+
"model.layers.26.self_attn.o_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1783 |
+
"model.layers.26.self_attn.qkv_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1784 |
+
"model.layers.26.self_attn.qkv_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1785 |
+
"model.layers.26.self_attn.qkv_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1786 |
+
"model.layers.26.self_attn.qkv_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1787 |
+
"model.layers.26.self_attn.qkv_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1788 |
"model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1789 |
"model.layers.27.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1790 |
"model.layers.27.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
|
|
1797 |
"model.layers.27.mlp.gate_up_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1798 |
"model.layers.27.mlp.gate_up_proj.lora_B.vision.weight": "model-00003-of-00003.safetensors",
|
1799 |
"model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
1800 |
+
"model.layers.27.self_attn.o_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1801 |
+
"model.layers.27.self_attn.o_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1802 |
+
"model.layers.27.self_attn.o_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1803 |
+
"model.layers.27.self_attn.o_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1804 |
+
"model.layers.27.self_attn.o_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1805 |
+
"model.layers.27.self_attn.qkv_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1806 |
+
"model.layers.27.self_attn.qkv_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1807 |
+
"model.layers.27.self_attn.qkv_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1808 |
+
"model.layers.27.self_attn.qkv_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1809 |
+
"model.layers.27.self_attn.qkv_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1810 |
"model.layers.28.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1811 |
"model.layers.28.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1812 |
"model.layers.28.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
|
|
1989 |
"model.layers.7.mlp.down_proj.lora_A.vision.weight": "model-00002-of-00003.safetensors",
|
1990 |
"model.layers.7.mlp.down_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1991 |
"model.layers.7.mlp.down_proj.lora_B.vision.weight": "model-00002-of-00003.safetensors",
|
1992 |
+
"model.layers.7.mlp.gate_up_proj.base_layer.weight": "model-00001-of-00003.safetensors",
|
1993 |
+
"model.layers.7.mlp.gate_up_proj.lora_A.speech.weight": "model-00001-of-00003.safetensors",
|
1994 |
+
"model.layers.7.mlp.gate_up_proj.lora_A.vision.weight": "model-00001-of-00003.safetensors",
|
1995 |
"model.layers.7.mlp.gate_up_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1996 |
+
"model.layers.7.mlp.gate_up_proj.lora_B.vision.weight": "model-00001-of-00003.safetensors",
|
1997 |
"model.layers.7.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
1998 |
"model.layers.7.self_attn.o_proj.base_layer.weight": "model-00001-of-00003.safetensors",
|
1999 |
+
"model.layers.7.self_attn.o_proj.lora_A.speech.weight": "model-00001-of-00003.safetensors",
|
2000 |
+
"model.layers.7.self_attn.o_proj.lora_A.vision.weight": "model-00001-of-00003.safetensors",
|
2001 |
+
"model.layers.7.self_attn.o_proj.lora_B.speech.weight": "model-00001-of-00003.safetensors",
|
2002 |
+
"model.layers.7.self_attn.o_proj.lora_B.vision.weight": "model-00001-of-00003.safetensors",
|
2003 |
+
"model.layers.7.self_attn.qkv_proj.base_layer.weight": "model-00001-of-00003.safetensors",
|
2004 |
+
"model.layers.7.self_attn.qkv_proj.lora_A.speech.weight": "model-00001-of-00003.safetensors",
|
2005 |
+
"model.layers.7.self_attn.qkv_proj.lora_A.vision.weight": "model-00001-of-00003.safetensors",
|
2006 |
+
"model.layers.7.self_attn.qkv_proj.lora_B.speech.weight": "model-00001-of-00003.safetensors",
|
2007 |
+
"model.layers.7.self_attn.qkv_proj.lora_B.vision.weight": "model-00001-of-00003.safetensors",
|
2008 |
"model.layers.8.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
2009 |
"model.layers.8.mlp.down_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
2010 |
"model.layers.8.mlp.down_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|