Upload MusicgenMelodyForConditionalGeneration
Browse files- config.json +2 -2
- model-00001-of-00002.safetensors +2 -2
- model-00002-of-00002.safetensors +2 -2
- model.safetensors.index.json +13 -5
config.json
CHANGED
@@ -105,7 +105,7 @@
|
|
105 |
"add_cross_attention": false,
|
106 |
"architectures": null,
|
107 |
"attention_dropout": 0.0,
|
108 |
-
"audio_channels":
|
109 |
"bad_words_ids": null,
|
110 |
"begin_suppress_tokens": null,
|
111 |
"bos_token_id": 2048,
|
@@ -145,7 +145,7 @@
|
|
145 |
"num_attention_heads": 24,
|
146 |
"num_beam_groups": 1,
|
147 |
"num_beams": 1,
|
148 |
-
"num_codebooks":
|
149 |
"num_hidden_layers": 48,
|
150 |
"num_return_sequences": 1,
|
151 |
"output_attentions": false,
|
|
|
105 |
"add_cross_attention": false,
|
106 |
"architectures": null,
|
107 |
"attention_dropout": 0.0,
|
108 |
+
"audio_channels": 2,
|
109 |
"bad_words_ids": null,
|
110 |
"begin_suppress_tokens": null,
|
111 |
"bos_token_id": 2048,
|
|
|
145 |
"num_attention_heads": 24,
|
146 |
"num_beam_groups": 1,
|
147 |
"num_beams": 1,
|
148 |
+
"num_codebooks": 8,
|
149 |
"num_hidden_layers": 48,
|
150 |
"num_return_sequences": 1,
|
151 |
"output_attentions": false,
|
model-00001-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e20ca49644e26560e90ccf54557852c59c13e8863337bd1562fd906ef0400515
|
3 |
+
size 4997739056
|
model-00002-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:95932e81bf5b2b02777a245a42dc65b9ab13362460ae77f7a1c3c0b50737e3c6
|
3 |
+
size 1332599632
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"audio_enc_to_dec_proj.bias": "model-00002-of-00002.safetensors",
|
@@ -125,11 +125,19 @@
|
|
125 |
"decoder.lm_heads.1.weight": "model-00002-of-00002.safetensors",
|
126 |
"decoder.lm_heads.2.weight": "model-00002-of-00002.safetensors",
|
127 |
"decoder.lm_heads.3.weight": "model-00002-of-00002.safetensors",
|
|
|
|
|
|
|
|
|
128 |
"decoder.model.decoder.embed_positions.weights": "model-00001-of-00002.safetensors",
|
129 |
"decoder.model.decoder.embed_tokens.0.weight": "model-00001-of-00002.safetensors",
|
130 |
"decoder.model.decoder.embed_tokens.1.weight": "model-00001-of-00002.safetensors",
|
131 |
"decoder.model.decoder.embed_tokens.2.weight": "model-00001-of-00002.safetensors",
|
132 |
"decoder.model.decoder.embed_tokens.3.weight": "model-00001-of-00002.safetensors",
|
|
|
|
|
|
|
|
|
133 |
"decoder.model.decoder.layer_norm.bias": "model-00002-of-00002.safetensors",
|
134 |
"decoder.model.decoder.layer_norm.weight": "model-00002-of-00002.safetensors",
|
135 |
"decoder.model.decoder.layers.0.fc1.weight": "model-00001-of-00002.safetensors",
|
@@ -447,11 +455,11 @@
|
|
447 |
"decoder.model.decoder.layers.37.final_layer_norm.bias": "model-00002-of-00002.safetensors",
|
448 |
"decoder.model.decoder.layers.37.final_layer_norm.weight": "model-00002-of-00002.safetensors",
|
449 |
"decoder.model.decoder.layers.37.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
450 |
-
"decoder.model.decoder.layers.37.self_attn.out_proj.weight": "model-
|
451 |
-
"decoder.model.decoder.layers.37.self_attn.q_proj.weight": "model-
|
452 |
"decoder.model.decoder.layers.37.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
453 |
-
"decoder.model.decoder.layers.37.self_attn_layer_norm.bias": "model-
|
454 |
-
"decoder.model.decoder.layers.37.self_attn_layer_norm.weight": "model-
|
455 |
"decoder.model.decoder.layers.38.fc1.weight": "model-00002-of-00002.safetensors",
|
456 |
"decoder.model.decoder.layers.38.fc2.weight": "model-00002-of-00002.safetensors",
|
457 |
"decoder.model.decoder.layers.38.final_layer_norm.bias": "model-00002-of-00002.safetensors",
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 6330247448
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"audio_enc_to_dec_proj.bias": "model-00002-of-00002.safetensors",
|
|
|
125 |
"decoder.lm_heads.1.weight": "model-00002-of-00002.safetensors",
|
126 |
"decoder.lm_heads.2.weight": "model-00002-of-00002.safetensors",
|
127 |
"decoder.lm_heads.3.weight": "model-00002-of-00002.safetensors",
|
128 |
+
"decoder.lm_heads.4.weight": "model-00002-of-00002.safetensors",
|
129 |
+
"decoder.lm_heads.5.weight": "model-00002-of-00002.safetensors",
|
130 |
+
"decoder.lm_heads.6.weight": "model-00002-of-00002.safetensors",
|
131 |
+
"decoder.lm_heads.7.weight": "model-00002-of-00002.safetensors",
|
132 |
"decoder.model.decoder.embed_positions.weights": "model-00001-of-00002.safetensors",
|
133 |
"decoder.model.decoder.embed_tokens.0.weight": "model-00001-of-00002.safetensors",
|
134 |
"decoder.model.decoder.embed_tokens.1.weight": "model-00001-of-00002.safetensors",
|
135 |
"decoder.model.decoder.embed_tokens.2.weight": "model-00001-of-00002.safetensors",
|
136 |
"decoder.model.decoder.embed_tokens.3.weight": "model-00001-of-00002.safetensors",
|
137 |
+
"decoder.model.decoder.embed_tokens.4.weight": "model-00001-of-00002.safetensors",
|
138 |
+
"decoder.model.decoder.embed_tokens.5.weight": "model-00001-of-00002.safetensors",
|
139 |
+
"decoder.model.decoder.embed_tokens.6.weight": "model-00001-of-00002.safetensors",
|
140 |
+
"decoder.model.decoder.embed_tokens.7.weight": "model-00001-of-00002.safetensors",
|
141 |
"decoder.model.decoder.layer_norm.bias": "model-00002-of-00002.safetensors",
|
142 |
"decoder.model.decoder.layer_norm.weight": "model-00002-of-00002.safetensors",
|
143 |
"decoder.model.decoder.layers.0.fc1.weight": "model-00001-of-00002.safetensors",
|
|
|
455 |
"decoder.model.decoder.layers.37.final_layer_norm.bias": "model-00002-of-00002.safetensors",
|
456 |
"decoder.model.decoder.layers.37.final_layer_norm.weight": "model-00002-of-00002.safetensors",
|
457 |
"decoder.model.decoder.layers.37.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
|
458 |
+
"decoder.model.decoder.layers.37.self_attn.out_proj.weight": "model-00002-of-00002.safetensors",
|
459 |
+
"decoder.model.decoder.layers.37.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
|
460 |
"decoder.model.decoder.layers.37.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
|
461 |
+
"decoder.model.decoder.layers.37.self_attn_layer_norm.bias": "model-00002-of-00002.safetensors",
|
462 |
+
"decoder.model.decoder.layers.37.self_attn_layer_norm.weight": "model-00002-of-00002.safetensors",
|
463 |
"decoder.model.decoder.layers.38.fc1.weight": "model-00002-of-00002.safetensors",
|
464 |
"decoder.model.decoder.layers.38.fc2.weight": "model-00002-of-00002.safetensors",
|
465 |
"decoder.model.decoder.layers.38.final_layer_norm.bias": "model-00002-of-00002.safetensors",
|