Update Speech Tower Training
Browse files- config.json +3 -3
- generation_config.json +1 -1
- model-00001-of-00003.safetensors +2 -2
- model-00002-of-00003.safetensors +2 -2
- model-00003-of-00003.safetensors +2 -2
- model.safetensors.index.json +142 -142
- processor_config.json +0 -3
- speech/adapter_model.safetensors +2 -2
config.json
CHANGED
@@ -45,8 +45,8 @@
|
|
45 |
},
|
46 |
"audio_token_index": 262143,
|
47 |
"auto_map": {
|
48 |
-
"AutoConfig": "configuration_gemma3mm.Gemma3MMConfig",
|
49 |
-
"AutoModel": "modeling_gemma3mm.Gemma3MMForConditionalGeneration"
|
50 |
},
|
51 |
"boa_token_index": 256001,
|
52 |
"boi_token_index": 255999,
|
@@ -98,7 +98,7 @@
|
|
98 |
"vocab_size": 262208
|
99 |
},
|
100 |
"torch_dtype": "bfloat16",
|
101 |
-
"transformers_version": "4.
|
102 |
"use_cache": false,
|
103 |
"vision_config": {
|
104 |
"attention_dropout": 0.0,
|
|
|
45 |
},
|
46 |
"audio_token_index": 262143,
|
47 |
"auto_map": {
|
48 |
+
"AutoConfig": "junnei/gemma-3-4b-it-speech--configuration_gemma3mm.Gemma3MMConfig",
|
49 |
+
"AutoModel": "junnei/gemma-3-4b-it-speech--modeling_gemma3mm.Gemma3MMForConditionalGeneration"
|
50 |
},
|
51 |
"boa_token_index": 256001,
|
52 |
"boi_token_index": 255999,
|
|
|
98 |
"vocab_size": 262208
|
99 |
},
|
100 |
"torch_dtype": "bfloat16",
|
101 |
+
"transformers_version": "4.51.0.dev0",
|
102 |
"use_cache": false,
|
103 |
"vision_config": {
|
104 |
"attention_dropout": 0.0,
|
generation_config.json
CHANGED
@@ -7,5 +7,5 @@
|
|
7 |
106
|
8 |
],
|
9 |
"pad_token_id": 0,
|
10 |
-
"transformers_version": "4.
|
11 |
}
|
|
|
7 |
106
|
8 |
],
|
9 |
"pad_token_id": 0,
|
10 |
+
"transformers_version": "4.51.0.dev0"
|
11 |
}
|
model-00001-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:063ff246586b3cccf3d28be470e42b79303bdc145d1192940e0cb98ca847d4f5
|
3 |
+
size 4947827632
|
model-00002-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f616857a2fbb5eb84de53e9583f5ddc75622e2ef912aab204b829f03a807fcdd
|
3 |
+
size 4987779168
|
model-00003-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:20497f36e13c3e7554613d9d4b25455efc058422718b9a871c2f246f14dac586
|
3 |
+
size 1949903376
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"audio_projector.0.bias": "model-00001-of-00003.safetensors",
|
@@ -1400,19 +1400,19 @@
|
|
1400 |
"language_model.model.base_model.model.layers.25.self_attn.v_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1401 |
"language_model.model.base_model.model.layers.25.self_attn.v_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1402 |
"language_model.model.base_model.model.layers.25.self_attn.v_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1403 |
-
"language_model.model.base_model.model.layers.26.input_layernorm.weight": "model-
|
1404 |
-
"language_model.model.base_model.model.layers.26.mlp.down_proj.base_layer.weight": "model-
|
1405 |
-
"language_model.model.base_model.model.layers.26.mlp.down_proj.lora_A.speech.weight": "model-
|
1406 |
-
"language_model.model.base_model.model.layers.26.mlp.down_proj.lora_B.speech.weight": "model-
|
1407 |
"language_model.model.base_model.model.layers.26.mlp.gate_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1408 |
"language_model.model.base_model.model.layers.26.mlp.gate_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1409 |
"language_model.model.base_model.model.layers.26.mlp.gate_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1410 |
-
"language_model.model.base_model.model.layers.26.mlp.up_proj.base_layer.weight": "model-
|
1411 |
-
"language_model.model.base_model.model.layers.26.mlp.up_proj.lora_A.speech.weight": "model-
|
1412 |
-
"language_model.model.base_model.model.layers.26.mlp.up_proj.lora_B.speech.weight": "model-
|
1413 |
-
"language_model.model.base_model.model.layers.26.post_attention_layernorm.weight": "model-
|
1414 |
-
"language_model.model.base_model.model.layers.26.post_feedforward_layernorm.weight": "model-
|
1415 |
-
"language_model.model.base_model.model.layers.26.pre_feedforward_layernorm.weight": "model-
|
1416 |
"language_model.model.base_model.model.layers.26.self_attn.k_norm.weight": "model-00002-of-00003.safetensors",
|
1417 |
"language_model.model.base_model.model.layers.26.self_attn.k_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1418 |
"language_model.model.base_model.model.layers.26.self_attn.k_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
@@ -1427,87 +1427,87 @@
|
|
1427 |
"language_model.model.base_model.model.layers.26.self_attn.v_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1428 |
"language_model.model.base_model.model.layers.26.self_attn.v_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1429 |
"language_model.model.base_model.model.layers.26.self_attn.v_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1430 |
-
"language_model.model.base_model.model.layers.27.input_layernorm.weight": "model-
|
1431 |
-
"language_model.model.base_model.model.layers.27.mlp.down_proj.base_layer.weight": "model-
|
1432 |
-
"language_model.model.base_model.model.layers.27.mlp.down_proj.lora_A.speech.weight": "model-
|
1433 |
-
"language_model.model.base_model.model.layers.27.mlp.down_proj.lora_B.speech.weight": "model-
|
1434 |
-
"language_model.model.base_model.model.layers.27.mlp.gate_proj.base_layer.weight": "model-
|
1435 |
-
"language_model.model.base_model.model.layers.27.mlp.gate_proj.lora_A.speech.weight": "model-
|
1436 |
-
"language_model.model.base_model.model.layers.27.mlp.gate_proj.lora_B.speech.weight": "model-
|
1437 |
-
"language_model.model.base_model.model.layers.27.mlp.up_proj.base_layer.weight": "model-
|
1438 |
-
"language_model.model.base_model.model.layers.27.mlp.up_proj.lora_A.speech.weight": "model-
|
1439 |
-
"language_model.model.base_model.model.layers.27.mlp.up_proj.lora_B.speech.weight": "model-
|
1440 |
-
"language_model.model.base_model.model.layers.27.post_attention_layernorm.weight": "model-
|
1441 |
-
"language_model.model.base_model.model.layers.27.post_feedforward_layernorm.weight": "model-
|
1442 |
-
"language_model.model.base_model.model.layers.27.pre_feedforward_layernorm.weight": "model-
|
1443 |
-
"language_model.model.base_model.model.layers.27.self_attn.k_norm.weight": "model-
|
1444 |
-
"language_model.model.base_model.model.layers.27.self_attn.k_proj.base_layer.weight": "model-
|
1445 |
-
"language_model.model.base_model.model.layers.27.self_attn.k_proj.lora_A.speech.weight": "model-
|
1446 |
-
"language_model.model.base_model.model.layers.27.self_attn.k_proj.lora_B.speech.weight": "model-
|
1447 |
-
"language_model.model.base_model.model.layers.27.self_attn.o_proj.base_layer.weight": "model-
|
1448 |
-
"language_model.model.base_model.model.layers.27.self_attn.o_proj.lora_A.speech.weight": "model-
|
1449 |
-
"language_model.model.base_model.model.layers.27.self_attn.o_proj.lora_B.speech.weight": "model-
|
1450 |
-
"language_model.model.base_model.model.layers.27.self_attn.q_norm.weight": "model-
|
1451 |
-
"language_model.model.base_model.model.layers.27.self_attn.q_proj.base_layer.weight": "model-
|
1452 |
-
"language_model.model.base_model.model.layers.27.self_attn.q_proj.lora_A.speech.weight": "model-
|
1453 |
-
"language_model.model.base_model.model.layers.27.self_attn.q_proj.lora_B.speech.weight": "model-
|
1454 |
-
"language_model.model.base_model.model.layers.27.self_attn.v_proj.base_layer.weight": "model-
|
1455 |
-
"language_model.model.base_model.model.layers.27.self_attn.v_proj.lora_A.speech.weight": "model-
|
1456 |
-
"language_model.model.base_model.model.layers.27.self_attn.v_proj.lora_B.speech.weight": "model-
|
1457 |
-
"language_model.model.base_model.model.layers.28.input_layernorm.weight": "model-
|
1458 |
-
"language_model.model.base_model.model.layers.28.mlp.down_proj.base_layer.weight": "model-
|
1459 |
-
"language_model.model.base_model.model.layers.28.mlp.down_proj.lora_A.speech.weight": "model-
|
1460 |
-
"language_model.model.base_model.model.layers.28.mlp.down_proj.lora_B.speech.weight": "model-
|
1461 |
-
"language_model.model.base_model.model.layers.28.mlp.gate_proj.base_layer.weight": "model-
|
1462 |
-
"language_model.model.base_model.model.layers.28.mlp.gate_proj.lora_A.speech.weight": "model-
|
1463 |
-
"language_model.model.base_model.model.layers.28.mlp.gate_proj.lora_B.speech.weight": "model-
|
1464 |
-
"language_model.model.base_model.model.layers.28.mlp.up_proj.base_layer.weight": "model-
|
1465 |
-
"language_model.model.base_model.model.layers.28.mlp.up_proj.lora_A.speech.weight": "model-
|
1466 |
-
"language_model.model.base_model.model.layers.28.mlp.up_proj.lora_B.speech.weight": "model-
|
1467 |
-
"language_model.model.base_model.model.layers.28.post_attention_layernorm.weight": "model-
|
1468 |
-
"language_model.model.base_model.model.layers.28.post_feedforward_layernorm.weight": "model-
|
1469 |
-
"language_model.model.base_model.model.layers.28.pre_feedforward_layernorm.weight": "model-
|
1470 |
-
"language_model.model.base_model.model.layers.28.self_attn.k_norm.weight": "model-
|
1471 |
-
"language_model.model.base_model.model.layers.28.self_attn.k_proj.base_layer.weight": "model-
|
1472 |
-
"language_model.model.base_model.model.layers.28.self_attn.k_proj.lora_A.speech.weight": "model-
|
1473 |
-
"language_model.model.base_model.model.layers.28.self_attn.k_proj.lora_B.speech.weight": "model-
|
1474 |
-
"language_model.model.base_model.model.layers.28.self_attn.o_proj.base_layer.weight": "model-
|
1475 |
-
"language_model.model.base_model.model.layers.28.self_attn.o_proj.lora_A.speech.weight": "model-
|
1476 |
-
"language_model.model.base_model.model.layers.28.self_attn.o_proj.lora_B.speech.weight": "model-
|
1477 |
-
"language_model.model.base_model.model.layers.28.self_attn.q_norm.weight": "model-
|
1478 |
-
"language_model.model.base_model.model.layers.28.self_attn.q_proj.base_layer.weight": "model-
|
1479 |
-
"language_model.model.base_model.model.layers.28.self_attn.q_proj.lora_A.speech.weight": "model-
|
1480 |
-
"language_model.model.base_model.model.layers.28.self_attn.q_proj.lora_B.speech.weight": "model-
|
1481 |
-
"language_model.model.base_model.model.layers.28.self_attn.v_proj.base_layer.weight": "model-
|
1482 |
-
"language_model.model.base_model.model.layers.28.self_attn.v_proj.lora_A.speech.weight": "model-
|
1483 |
-
"language_model.model.base_model.model.layers.28.self_attn.v_proj.lora_B.speech.weight": "model-
|
1484 |
-
"language_model.model.base_model.model.layers.29.input_layernorm.weight": "model-
|
1485 |
-
"language_model.model.base_model.model.layers.29.mlp.down_proj.base_layer.weight": "model-
|
1486 |
-
"language_model.model.base_model.model.layers.29.mlp.down_proj.lora_A.speech.weight": "model-
|
1487 |
-
"language_model.model.base_model.model.layers.29.mlp.down_proj.lora_B.speech.weight": "model-
|
1488 |
-
"language_model.model.base_model.model.layers.29.mlp.gate_proj.base_layer.weight": "model-
|
1489 |
-
"language_model.model.base_model.model.layers.29.mlp.gate_proj.lora_A.speech.weight": "model-
|
1490 |
-
"language_model.model.base_model.model.layers.29.mlp.gate_proj.lora_B.speech.weight": "model-
|
1491 |
-
"language_model.model.base_model.model.layers.29.mlp.up_proj.base_layer.weight": "model-
|
1492 |
-
"language_model.model.base_model.model.layers.29.mlp.up_proj.lora_A.speech.weight": "model-
|
1493 |
-
"language_model.model.base_model.model.layers.29.mlp.up_proj.lora_B.speech.weight": "model-
|
1494 |
-
"language_model.model.base_model.model.layers.29.post_attention_layernorm.weight": "model-
|
1495 |
-
"language_model.model.base_model.model.layers.29.post_feedforward_layernorm.weight": "model-
|
1496 |
-
"language_model.model.base_model.model.layers.29.pre_feedforward_layernorm.weight": "model-
|
1497 |
-
"language_model.model.base_model.model.layers.29.self_attn.k_norm.weight": "model-
|
1498 |
-
"language_model.model.base_model.model.layers.29.self_attn.k_proj.base_layer.weight": "model-
|
1499 |
-
"language_model.model.base_model.model.layers.29.self_attn.k_proj.lora_A.speech.weight": "model-
|
1500 |
-
"language_model.model.base_model.model.layers.29.self_attn.k_proj.lora_B.speech.weight": "model-
|
1501 |
-
"language_model.model.base_model.model.layers.29.self_attn.o_proj.base_layer.weight": "model-
|
1502 |
-
"language_model.model.base_model.model.layers.29.self_attn.o_proj.lora_A.speech.weight": "model-
|
1503 |
-
"language_model.model.base_model.model.layers.29.self_attn.o_proj.lora_B.speech.weight": "model-
|
1504 |
-
"language_model.model.base_model.model.layers.29.self_attn.q_norm.weight": "model-
|
1505 |
-
"language_model.model.base_model.model.layers.29.self_attn.q_proj.base_layer.weight": "model-
|
1506 |
-
"language_model.model.base_model.model.layers.29.self_attn.q_proj.lora_A.speech.weight": "model-
|
1507 |
-
"language_model.model.base_model.model.layers.29.self_attn.q_proj.lora_B.speech.weight": "model-
|
1508 |
-
"language_model.model.base_model.model.layers.29.self_attn.v_proj.base_layer.weight": "model-
|
1509 |
-
"language_model.model.base_model.model.layers.29.self_attn.v_proj.lora_A.speech.weight": "model-
|
1510 |
-
"language_model.model.base_model.model.layers.29.self_attn.v_proj.lora_B.speech.weight": "model-
|
1511 |
"language_model.model.base_model.model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
1512 |
"language_model.model.base_model.model.layers.3.mlp.down_proj.base_layer.weight": "model-00001-of-00003.safetensors",
|
1513 |
"language_model.model.base_model.model.layers.3.mlp.down_proj.lora_A.speech.weight": "model-00001-of-00003.safetensors",
|
@@ -1539,29 +1539,29 @@
|
|
1539 |
"language_model.model.base_model.model.layers.30.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1540 |
"language_model.model.base_model.model.layers.30.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1541 |
"language_model.model.base_model.model.layers.30.mlp.down_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1542 |
-
"language_model.model.base_model.model.layers.30.mlp.gate_proj.base_layer.weight": "model-
|
1543 |
-
"language_model.model.base_model.model.layers.30.mlp.gate_proj.lora_A.speech.weight": "model-
|
1544 |
-
"language_model.model.base_model.model.layers.30.mlp.gate_proj.lora_B.speech.weight": "model-
|
1545 |
-
"language_model.model.base_model.model.layers.30.mlp.up_proj.base_layer.weight": "model-
|
1546 |
-
"language_model.model.base_model.model.layers.30.mlp.up_proj.lora_A.speech.weight": "model-
|
1547 |
-
"language_model.model.base_model.model.layers.30.mlp.up_proj.lora_B.speech.weight": "model-
|
1548 |
"language_model.model.base_model.model.layers.30.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
1549 |
"language_model.model.base_model.model.layers.30.post_feedforward_layernorm.weight": "model-00003-of-00003.safetensors",
|
1550 |
"language_model.model.base_model.model.layers.30.pre_feedforward_layernorm.weight": "model-00003-of-00003.safetensors",
|
1551 |
-
"language_model.model.base_model.model.layers.30.self_attn.k_norm.weight": "model-
|
1552 |
-
"language_model.model.base_model.model.layers.30.self_attn.k_proj.base_layer.weight": "model-
|
1553 |
-
"language_model.model.base_model.model.layers.30.self_attn.k_proj.lora_A.speech.weight": "model-
|
1554 |
-
"language_model.model.base_model.model.layers.30.self_attn.k_proj.lora_B.speech.weight": "model-
|
1555 |
-
"language_model.model.base_model.model.layers.30.self_attn.o_proj.base_layer.weight": "model-
|
1556 |
-
"language_model.model.base_model.model.layers.30.self_attn.o_proj.lora_A.speech.weight": "model-
|
1557 |
-
"language_model.model.base_model.model.layers.30.self_attn.o_proj.lora_B.speech.weight": "model-
|
1558 |
-
"language_model.model.base_model.model.layers.30.self_attn.q_norm.weight": "model-
|
1559 |
-
"language_model.model.base_model.model.layers.30.self_attn.q_proj.base_layer.weight": "model-
|
1560 |
-
"language_model.model.base_model.model.layers.30.self_attn.q_proj.lora_A.speech.weight": "model-
|
1561 |
-
"language_model.model.base_model.model.layers.30.self_attn.q_proj.lora_B.speech.weight": "model-
|
1562 |
-
"language_model.model.base_model.model.layers.30.self_attn.v_proj.base_layer.weight": "model-
|
1563 |
-
"language_model.model.base_model.model.layers.30.self_attn.v_proj.lora_A.speech.weight": "model-
|
1564 |
-
"language_model.model.base_model.model.layers.30.self_attn.v_proj.lora_B.speech.weight": "model-
|
1565 |
"language_model.model.base_model.model.layers.31.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1566 |
"language_model.model.base_model.model.layers.31.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1567 |
"language_model.model.base_model.model.layers.31.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
@@ -1724,19 +1724,19 @@
|
|
1724 |
"language_model.model.base_model.model.layers.6.self_attn.v_proj.base_layer.weight": "model-00001-of-00003.safetensors",
|
1725 |
"language_model.model.base_model.model.layers.6.self_attn.v_proj.lora_A.speech.weight": "model-00001-of-00003.safetensors",
|
1726 |
"language_model.model.base_model.model.layers.6.self_attn.v_proj.lora_B.speech.weight": "model-00001-of-00003.safetensors",
|
1727 |
-
"language_model.model.base_model.model.layers.7.input_layernorm.weight": "model-
|
1728 |
-
"language_model.model.base_model.model.layers.7.mlp.down_proj.base_layer.weight": "model-
|
1729 |
-
"language_model.model.base_model.model.layers.7.mlp.down_proj.lora_A.speech.weight": "model-
|
1730 |
-
"language_model.model.base_model.model.layers.7.mlp.down_proj.lora_B.speech.weight": "model-
|
1731 |
-
"language_model.model.base_model.model.layers.7.mlp.gate_proj.base_layer.weight": "model-
|
1732 |
-
"language_model.model.base_model.model.layers.7.mlp.gate_proj.lora_A.speech.weight": "model-
|
1733 |
-
"language_model.model.base_model.model.layers.7.mlp.gate_proj.lora_B.speech.weight": "model-
|
1734 |
-
"language_model.model.base_model.model.layers.7.mlp.up_proj.base_layer.weight": "model-
|
1735 |
-
"language_model.model.base_model.model.layers.7.mlp.up_proj.lora_A.speech.weight": "model-
|
1736 |
-
"language_model.model.base_model.model.layers.7.mlp.up_proj.lora_B.speech.weight": "model-
|
1737 |
-
"language_model.model.base_model.model.layers.7.post_attention_layernorm.weight": "model-
|
1738 |
-
"language_model.model.base_model.model.layers.7.post_feedforward_layernorm.weight": "model-
|
1739 |
-
"language_model.model.base_model.model.layers.7.pre_feedforward_layernorm.weight": "model-
|
1740 |
"language_model.model.base_model.model.layers.7.self_attn.k_norm.weight": "model-00001-of-00003.safetensors",
|
1741 |
"language_model.model.base_model.model.layers.7.self_attn.k_proj.base_layer.weight": "model-00001-of-00003.safetensors",
|
1742 |
"language_model.model.base_model.model.layers.7.self_attn.k_proj.lora_A.speech.weight": "model-00001-of-00003.safetensors",
|
@@ -1755,29 +1755,29 @@
|
|
1755 |
"language_model.model.base_model.model.layers.8.mlp.down_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1756 |
"language_model.model.base_model.model.layers.8.mlp.down_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1757 |
"language_model.model.base_model.model.layers.8.mlp.down_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1758 |
-
"language_model.model.base_model.model.layers.8.mlp.gate_proj.base_layer.weight": "model-
|
1759 |
-
"language_model.model.base_model.model.layers.8.mlp.gate_proj.lora_A.speech.weight": "model-
|
1760 |
-
"language_model.model.base_model.model.layers.8.mlp.gate_proj.lora_B.speech.weight": "model-
|
1761 |
"language_model.model.base_model.model.layers.8.mlp.up_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1762 |
"language_model.model.base_model.model.layers.8.mlp.up_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1763 |
"language_model.model.base_model.model.layers.8.mlp.up_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1764 |
"language_model.model.base_model.model.layers.8.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
1765 |
"language_model.model.base_model.model.layers.8.post_feedforward_layernorm.weight": "model-00002-of-00003.safetensors",
|
1766 |
"language_model.model.base_model.model.layers.8.pre_feedforward_layernorm.weight": "model-00002-of-00003.safetensors",
|
1767 |
-
"language_model.model.base_model.model.layers.8.self_attn.k_norm.weight": "model-
|
1768 |
-
"language_model.model.base_model.model.layers.8.self_attn.k_proj.base_layer.weight": "model-
|
1769 |
-
"language_model.model.base_model.model.layers.8.self_attn.k_proj.lora_A.speech.weight": "model-
|
1770 |
-
"language_model.model.base_model.model.layers.8.self_attn.k_proj.lora_B.speech.weight": "model-
|
1771 |
-
"language_model.model.base_model.model.layers.8.self_attn.o_proj.base_layer.weight": "model-
|
1772 |
-
"language_model.model.base_model.model.layers.8.self_attn.o_proj.lora_A.speech.weight": "model-
|
1773 |
-
"language_model.model.base_model.model.layers.8.self_attn.o_proj.lora_B.speech.weight": "model-
|
1774 |
-
"language_model.model.base_model.model.layers.8.self_attn.q_norm.weight": "model-
|
1775 |
-
"language_model.model.base_model.model.layers.8.self_attn.q_proj.base_layer.weight": "model-
|
1776 |
-
"language_model.model.base_model.model.layers.8.self_attn.q_proj.lora_A.speech.weight": "model-
|
1777 |
-
"language_model.model.base_model.model.layers.8.self_attn.q_proj.lora_B.speech.weight": "model-
|
1778 |
-
"language_model.model.base_model.model.layers.8.self_attn.v_proj.base_layer.weight": "model-
|
1779 |
-
"language_model.model.base_model.model.layers.8.self_attn.v_proj.lora_A.speech.weight": "model-
|
1780 |
-
"language_model.model.base_model.model.layers.8.self_attn.v_proj.lora_B.speech.weight": "model-
|
1781 |
"language_model.model.base_model.model.layers.9.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
1782 |
"language_model.model.base_model.model.layers.9.mlp.down_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1783 |
"language_model.model.base_model.model.layers.9.mlp.down_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 11885195552
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"audio_projector.0.bias": "model-00001-of-00003.safetensors",
|
|
|
1400 |
"language_model.model.base_model.model.layers.25.self_attn.v_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1401 |
"language_model.model.base_model.model.layers.25.self_attn.v_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1402 |
"language_model.model.base_model.model.layers.25.self_attn.v_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1403 |
+
"language_model.model.base_model.model.layers.26.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1404 |
+
"language_model.model.base_model.model.layers.26.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1405 |
+
"language_model.model.base_model.model.layers.26.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1406 |
+
"language_model.model.base_model.model.layers.26.mlp.down_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1407 |
"language_model.model.base_model.model.layers.26.mlp.gate_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1408 |
"language_model.model.base_model.model.layers.26.mlp.gate_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1409 |
"language_model.model.base_model.model.layers.26.mlp.gate_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1410 |
+
"language_model.model.base_model.model.layers.26.mlp.up_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1411 |
+
"language_model.model.base_model.model.layers.26.mlp.up_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1412 |
+
"language_model.model.base_model.model.layers.26.mlp.up_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1413 |
+
"language_model.model.base_model.model.layers.26.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
1414 |
+
"language_model.model.base_model.model.layers.26.post_feedforward_layernorm.weight": "model-00003-of-00003.safetensors",
|
1415 |
+
"language_model.model.base_model.model.layers.26.pre_feedforward_layernorm.weight": "model-00003-of-00003.safetensors",
|
1416 |
"language_model.model.base_model.model.layers.26.self_attn.k_norm.weight": "model-00002-of-00003.safetensors",
|
1417 |
"language_model.model.base_model.model.layers.26.self_attn.k_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1418 |
"language_model.model.base_model.model.layers.26.self_attn.k_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
|
|
1427 |
"language_model.model.base_model.model.layers.26.self_attn.v_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1428 |
"language_model.model.base_model.model.layers.26.self_attn.v_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1429 |
"language_model.model.base_model.model.layers.26.self_attn.v_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1430 |
+
"language_model.model.base_model.model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1431 |
+
"language_model.model.base_model.model.layers.27.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1432 |
+
"language_model.model.base_model.model.layers.27.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1433 |
+
"language_model.model.base_model.model.layers.27.mlp.down_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1434 |
+
"language_model.model.base_model.model.layers.27.mlp.gate_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1435 |
+
"language_model.model.base_model.model.layers.27.mlp.gate_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1436 |
+
"language_model.model.base_model.model.layers.27.mlp.gate_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1437 |
+
"language_model.model.base_model.model.layers.27.mlp.up_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1438 |
+
"language_model.model.base_model.model.layers.27.mlp.up_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1439 |
+
"language_model.model.base_model.model.layers.27.mlp.up_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1440 |
+
"language_model.model.base_model.model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
1441 |
+
"language_model.model.base_model.model.layers.27.post_feedforward_layernorm.weight": "model-00003-of-00003.safetensors",
|
1442 |
+
"language_model.model.base_model.model.layers.27.pre_feedforward_layernorm.weight": "model-00003-of-00003.safetensors",
|
1443 |
+
"language_model.model.base_model.model.layers.27.self_attn.k_norm.weight": "model-00003-of-00003.safetensors",
|
1444 |
+
"language_model.model.base_model.model.layers.27.self_attn.k_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1445 |
+
"language_model.model.base_model.model.layers.27.self_attn.k_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1446 |
+
"language_model.model.base_model.model.layers.27.self_attn.k_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1447 |
+
"language_model.model.base_model.model.layers.27.self_attn.o_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1448 |
+
"language_model.model.base_model.model.layers.27.self_attn.o_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1449 |
+
"language_model.model.base_model.model.layers.27.self_attn.o_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1450 |
+
"language_model.model.base_model.model.layers.27.self_attn.q_norm.weight": "model-00003-of-00003.safetensors",
|
1451 |
+
"language_model.model.base_model.model.layers.27.self_attn.q_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1452 |
+
"language_model.model.base_model.model.layers.27.self_attn.q_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1453 |
+
"language_model.model.base_model.model.layers.27.self_attn.q_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1454 |
+
"language_model.model.base_model.model.layers.27.self_attn.v_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1455 |
+
"language_model.model.base_model.model.layers.27.self_attn.v_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1456 |
+
"language_model.model.base_model.model.layers.27.self_attn.v_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1457 |
+
"language_model.model.base_model.model.layers.28.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1458 |
+
"language_model.model.base_model.model.layers.28.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1459 |
+
"language_model.model.base_model.model.layers.28.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1460 |
+
"language_model.model.base_model.model.layers.28.mlp.down_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1461 |
+
"language_model.model.base_model.model.layers.28.mlp.gate_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1462 |
+
"language_model.model.base_model.model.layers.28.mlp.gate_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1463 |
+
"language_model.model.base_model.model.layers.28.mlp.gate_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1464 |
+
"language_model.model.base_model.model.layers.28.mlp.up_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1465 |
+
"language_model.model.base_model.model.layers.28.mlp.up_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1466 |
+
"language_model.model.base_model.model.layers.28.mlp.up_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1467 |
+
"language_model.model.base_model.model.layers.28.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
1468 |
+
"language_model.model.base_model.model.layers.28.post_feedforward_layernorm.weight": "model-00003-of-00003.safetensors",
|
1469 |
+
"language_model.model.base_model.model.layers.28.pre_feedforward_layernorm.weight": "model-00003-of-00003.safetensors",
|
1470 |
+
"language_model.model.base_model.model.layers.28.self_attn.k_norm.weight": "model-00003-of-00003.safetensors",
|
1471 |
+
"language_model.model.base_model.model.layers.28.self_attn.k_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1472 |
+
"language_model.model.base_model.model.layers.28.self_attn.k_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1473 |
+
"language_model.model.base_model.model.layers.28.self_attn.k_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1474 |
+
"language_model.model.base_model.model.layers.28.self_attn.o_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1475 |
+
"language_model.model.base_model.model.layers.28.self_attn.o_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1476 |
+
"language_model.model.base_model.model.layers.28.self_attn.o_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1477 |
+
"language_model.model.base_model.model.layers.28.self_attn.q_norm.weight": "model-00003-of-00003.safetensors",
|
1478 |
+
"language_model.model.base_model.model.layers.28.self_attn.q_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1479 |
+
"language_model.model.base_model.model.layers.28.self_attn.q_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1480 |
+
"language_model.model.base_model.model.layers.28.self_attn.q_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1481 |
+
"language_model.model.base_model.model.layers.28.self_attn.v_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1482 |
+
"language_model.model.base_model.model.layers.28.self_attn.v_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1483 |
+
"language_model.model.base_model.model.layers.28.self_attn.v_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1484 |
+
"language_model.model.base_model.model.layers.29.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1485 |
+
"language_model.model.base_model.model.layers.29.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1486 |
+
"language_model.model.base_model.model.layers.29.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1487 |
+
"language_model.model.base_model.model.layers.29.mlp.down_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1488 |
+
"language_model.model.base_model.model.layers.29.mlp.gate_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1489 |
+
"language_model.model.base_model.model.layers.29.mlp.gate_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1490 |
+
"language_model.model.base_model.model.layers.29.mlp.gate_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1491 |
+
"language_model.model.base_model.model.layers.29.mlp.up_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1492 |
+
"language_model.model.base_model.model.layers.29.mlp.up_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1493 |
+
"language_model.model.base_model.model.layers.29.mlp.up_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1494 |
+
"language_model.model.base_model.model.layers.29.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
1495 |
+
"language_model.model.base_model.model.layers.29.post_feedforward_layernorm.weight": "model-00003-of-00003.safetensors",
|
1496 |
+
"language_model.model.base_model.model.layers.29.pre_feedforward_layernorm.weight": "model-00003-of-00003.safetensors",
|
1497 |
+
"language_model.model.base_model.model.layers.29.self_attn.k_norm.weight": "model-00003-of-00003.safetensors",
|
1498 |
+
"language_model.model.base_model.model.layers.29.self_attn.k_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1499 |
+
"language_model.model.base_model.model.layers.29.self_attn.k_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1500 |
+
"language_model.model.base_model.model.layers.29.self_attn.k_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1501 |
+
"language_model.model.base_model.model.layers.29.self_attn.o_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1502 |
+
"language_model.model.base_model.model.layers.29.self_attn.o_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1503 |
+
"language_model.model.base_model.model.layers.29.self_attn.o_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1504 |
+
"language_model.model.base_model.model.layers.29.self_attn.q_norm.weight": "model-00003-of-00003.safetensors",
|
1505 |
+
"language_model.model.base_model.model.layers.29.self_attn.q_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1506 |
+
"language_model.model.base_model.model.layers.29.self_attn.q_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1507 |
+
"language_model.model.base_model.model.layers.29.self_attn.q_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1508 |
+
"language_model.model.base_model.model.layers.29.self_attn.v_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1509 |
+
"language_model.model.base_model.model.layers.29.self_attn.v_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1510 |
+
"language_model.model.base_model.model.layers.29.self_attn.v_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1511 |
"language_model.model.base_model.model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
1512 |
"language_model.model.base_model.model.layers.3.mlp.down_proj.base_layer.weight": "model-00001-of-00003.safetensors",
|
1513 |
"language_model.model.base_model.model.layers.3.mlp.down_proj.lora_A.speech.weight": "model-00001-of-00003.safetensors",
|
|
|
1539 |
"language_model.model.base_model.model.layers.30.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1540 |
"language_model.model.base_model.model.layers.30.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1541 |
"language_model.model.base_model.model.layers.30.mlp.down_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1542 |
+
"language_model.model.base_model.model.layers.30.mlp.gate_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1543 |
+
"language_model.model.base_model.model.layers.30.mlp.gate_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1544 |
+
"language_model.model.base_model.model.layers.30.mlp.gate_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1545 |
+
"language_model.model.base_model.model.layers.30.mlp.up_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1546 |
+
"language_model.model.base_model.model.layers.30.mlp.up_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1547 |
+
"language_model.model.base_model.model.layers.30.mlp.up_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1548 |
"language_model.model.base_model.model.layers.30.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
1549 |
"language_model.model.base_model.model.layers.30.post_feedforward_layernorm.weight": "model-00003-of-00003.safetensors",
|
1550 |
"language_model.model.base_model.model.layers.30.pre_feedforward_layernorm.weight": "model-00003-of-00003.safetensors",
|
1551 |
+
"language_model.model.base_model.model.layers.30.self_attn.k_norm.weight": "model-00003-of-00003.safetensors",
|
1552 |
+
"language_model.model.base_model.model.layers.30.self_attn.k_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1553 |
+
"language_model.model.base_model.model.layers.30.self_attn.k_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1554 |
+
"language_model.model.base_model.model.layers.30.self_attn.k_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1555 |
+
"language_model.model.base_model.model.layers.30.self_attn.o_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1556 |
+
"language_model.model.base_model.model.layers.30.self_attn.o_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1557 |
+
"language_model.model.base_model.model.layers.30.self_attn.o_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1558 |
+
"language_model.model.base_model.model.layers.30.self_attn.q_norm.weight": "model-00003-of-00003.safetensors",
|
1559 |
+
"language_model.model.base_model.model.layers.30.self_attn.q_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1560 |
+
"language_model.model.base_model.model.layers.30.self_attn.q_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1561 |
+
"language_model.model.base_model.model.layers.30.self_attn.q_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1562 |
+
"language_model.model.base_model.model.layers.30.self_attn.v_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1563 |
+
"language_model.model.base_model.model.layers.30.self_attn.v_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
1564 |
+
"language_model.model.base_model.model.layers.30.self_attn.v_proj.lora_B.speech.weight": "model-00003-of-00003.safetensors",
|
1565 |
"language_model.model.base_model.model.layers.31.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
1566 |
"language_model.model.base_model.model.layers.31.mlp.down_proj.base_layer.weight": "model-00003-of-00003.safetensors",
|
1567 |
"language_model.model.base_model.model.layers.31.mlp.down_proj.lora_A.speech.weight": "model-00003-of-00003.safetensors",
|
|
|
1724 |
"language_model.model.base_model.model.layers.6.self_attn.v_proj.base_layer.weight": "model-00001-of-00003.safetensors",
|
1725 |
"language_model.model.base_model.model.layers.6.self_attn.v_proj.lora_A.speech.weight": "model-00001-of-00003.safetensors",
|
1726 |
"language_model.model.base_model.model.layers.6.self_attn.v_proj.lora_B.speech.weight": "model-00001-of-00003.safetensors",
|
1727 |
+
"language_model.model.base_model.model.layers.7.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
1728 |
+
"language_model.model.base_model.model.layers.7.mlp.down_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1729 |
+
"language_model.model.base_model.model.layers.7.mlp.down_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1730 |
+
"language_model.model.base_model.model.layers.7.mlp.down_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1731 |
+
"language_model.model.base_model.model.layers.7.mlp.gate_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1732 |
+
"language_model.model.base_model.model.layers.7.mlp.gate_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1733 |
+
"language_model.model.base_model.model.layers.7.mlp.gate_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1734 |
+
"language_model.model.base_model.model.layers.7.mlp.up_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1735 |
+
"language_model.model.base_model.model.layers.7.mlp.up_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1736 |
+
"language_model.model.base_model.model.layers.7.mlp.up_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1737 |
+
"language_model.model.base_model.model.layers.7.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
1738 |
+
"language_model.model.base_model.model.layers.7.post_feedforward_layernorm.weight": "model-00002-of-00003.safetensors",
|
1739 |
+
"language_model.model.base_model.model.layers.7.pre_feedforward_layernorm.weight": "model-00002-of-00003.safetensors",
|
1740 |
"language_model.model.base_model.model.layers.7.self_attn.k_norm.weight": "model-00001-of-00003.safetensors",
|
1741 |
"language_model.model.base_model.model.layers.7.self_attn.k_proj.base_layer.weight": "model-00001-of-00003.safetensors",
|
1742 |
"language_model.model.base_model.model.layers.7.self_attn.k_proj.lora_A.speech.weight": "model-00001-of-00003.safetensors",
|
|
|
1755 |
"language_model.model.base_model.model.layers.8.mlp.down_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1756 |
"language_model.model.base_model.model.layers.8.mlp.down_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1757 |
"language_model.model.base_model.model.layers.8.mlp.down_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1758 |
+
"language_model.model.base_model.model.layers.8.mlp.gate_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1759 |
+
"language_model.model.base_model.model.layers.8.mlp.gate_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1760 |
+
"language_model.model.base_model.model.layers.8.mlp.gate_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1761 |
"language_model.model.base_model.model.layers.8.mlp.up_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1762 |
"language_model.model.base_model.model.layers.8.mlp.up_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1763 |
"language_model.model.base_model.model.layers.8.mlp.up_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1764 |
"language_model.model.base_model.model.layers.8.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
1765 |
"language_model.model.base_model.model.layers.8.post_feedforward_layernorm.weight": "model-00002-of-00003.safetensors",
|
1766 |
"language_model.model.base_model.model.layers.8.pre_feedforward_layernorm.weight": "model-00002-of-00003.safetensors",
|
1767 |
+
"language_model.model.base_model.model.layers.8.self_attn.k_norm.weight": "model-00002-of-00003.safetensors",
|
1768 |
+
"language_model.model.base_model.model.layers.8.self_attn.k_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1769 |
+
"language_model.model.base_model.model.layers.8.self_attn.k_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1770 |
+
"language_model.model.base_model.model.layers.8.self_attn.k_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1771 |
+
"language_model.model.base_model.model.layers.8.self_attn.o_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1772 |
+
"language_model.model.base_model.model.layers.8.self_attn.o_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1773 |
+
"language_model.model.base_model.model.layers.8.self_attn.o_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1774 |
+
"language_model.model.base_model.model.layers.8.self_attn.q_norm.weight": "model-00002-of-00003.safetensors",
|
1775 |
+
"language_model.model.base_model.model.layers.8.self_attn.q_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1776 |
+
"language_model.model.base_model.model.layers.8.self_attn.q_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1777 |
+
"language_model.model.base_model.model.layers.8.self_attn.q_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1778 |
+
"language_model.model.base_model.model.layers.8.self_attn.v_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1779 |
+
"language_model.model.base_model.model.layers.8.self_attn.v_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
1780 |
+
"language_model.model.base_model.model.layers.8.self_attn.v_proj.lora_B.speech.weight": "model-00002-of-00003.safetensors",
|
1781 |
"language_model.model.base_model.model.layers.9.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
1782 |
"language_model.model.base_model.model.layers.9.mlp.down_proj.base_layer.weight": "model-00002-of-00003.safetensors",
|
1783 |
"language_model.model.base_model.model.layers.9.mlp.down_proj.lora_A.speech.weight": "model-00002-of-00003.safetensors",
|
processor_config.json
CHANGED
@@ -1,7 +1,4 @@
|
|
1 |
{
|
2 |
-
"auto_map": {
|
3 |
-
"AutoProcessor": "processing_gemma3mm.Gemma3MMProcessor"
|
4 |
-
},
|
5 |
"image_seq_length": 256,
|
6 |
"processor_class": "Gemma3MMProcessor"
|
7 |
}
|
|
|
1 |
{
|
|
|
|
|
|
|
2 |
"image_seq_length": 256,
|
3 |
"processor_class": "Gemma3MMProcessor"
|
4 |
}
|
speech/adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ac07e64184582b55efd1e8c9164e300f0290f27ddf3b9a9100539bf815f53fc2
|
3 |
+
size 2384262280
|