cwang621 commited on
Commit
c2db1e9
·
verified ·
1 Parent(s): 173a6fa

Upload 7 files

Browse files
config.json ADDED
@@ -0,0 +1,338 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "adapter_hidden_layers": 0,
3
+ "adapter_inner_dim": 512,
4
+ "adapter_type": "subsampler",
5
+ "architectures": [
6
+ "Blsp2Model"
7
+ ],
8
+ "conv_kernel_sizes": "5,5,5",
9
+ "kd_temperature": 2,
10
+ "lora_config": {
11
+ "auto_mapping": null,
12
+ "base_model_name_or_path": null,
13
+ "bias": "none",
14
+ "fan_in_fan_out": false,
15
+ "inference_mode": false,
16
+ "init_lora_weights": true,
17
+ "layers_pattern": null,
18
+ "layers_to_transform": null,
19
+ "lora_alpha": 16,
20
+ "lora_dropout": 0.05,
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 16,
24
+ "revision": null,
25
+ "target_modules": [
26
+ "c_attn",
27
+ "c_proj",
28
+ "w1",
29
+ "w2"
30
+ ],
31
+ "task_type": null
32
+ },
33
+ "lora_scope": "audio",
34
+ "num_emotions": 5,
35
+ "num_post_cif_layers": 4,
36
+ "num_pre_cif_layers": 4,
37
+ "qwen_config": {
38
+ "_name_or_path": "",
39
+ "add_cross_attention": false,
40
+ "architectures": [
41
+ "QWenLMHeadModel"
42
+ ],
43
+ "attn_dropout_prob": 0.0,
44
+ "auto_map": {
45
+ "AutoConfig": "configuration_qwen.QWenConfig",
46
+ "AutoModelForCausalLM": "modeling_qwen.QWenLMHeadModel"
47
+ },
48
+ "bad_words_ids": null,
49
+ "begin_suppress_tokens": null,
50
+ "bf16": false,
51
+ "bos_token_id": null,
52
+ "chunk_size_feed_forward": 0,
53
+ "cross_attention_hidden_size": null,
54
+ "decoder_start_token_id": null,
55
+ "diversity_penalty": 0.0,
56
+ "do_sample": false,
57
+ "early_stopping": false,
58
+ "emb_dropout_prob": 0.0,
59
+ "encoder_no_repeat_ngram_size": 0,
60
+ "eos_token_id": null,
61
+ "exponential_decay_length_penalty": null,
62
+ "finetuning_task": null,
63
+ "forced_bos_token_id": null,
64
+ "forced_eos_token_id": null,
65
+ "fp16": false,
66
+ "fp32": false,
67
+ "hidden_size": 4096,
68
+ "id2label": {
69
+ "0": "LABEL_0",
70
+ "1": "LABEL_1"
71
+ },
72
+ "initializer_range": 0.02,
73
+ "intermediate_size": 22016,
74
+ "is_decoder": false,
75
+ "is_encoder_decoder": false,
76
+ "kv_channels": 128,
77
+ "label2id": {
78
+ "LABEL_0": 0,
79
+ "LABEL_1": 1
80
+ },
81
+ "layer_norm_epsilon": 1e-06,
82
+ "length_penalty": 1.0,
83
+ "max_length": 20,
84
+ "max_position_embeddings": 8192,
85
+ "min_length": 0,
86
+ "model_type": "qwen",
87
+ "no_bias": true,
88
+ "no_repeat_ngram_size": 0,
89
+ "num_attention_heads": 32,
90
+ "num_beam_groups": 1,
91
+ "num_beams": 1,
92
+ "num_hidden_layers": 32,
93
+ "num_return_sequences": 1,
94
+ "onnx_safe": null,
95
+ "output_attentions": false,
96
+ "output_hidden_states": false,
97
+ "output_scores": false,
98
+ "pad_token_id": null,
99
+ "prefix": null,
100
+ "problem_type": null,
101
+ "pruned_heads": {},
102
+ "remove_invalid_values": false,
103
+ "repetition_penalty": 1.0,
104
+ "return_dict": true,
105
+ "return_dict_in_generate": false,
106
+ "rotary_emb_base": 10000,
107
+ "rotary_pct": 1.0,
108
+ "scale_attn_weights": true,
109
+ "sep_token_id": null,
110
+ "seq_length": 8192,
111
+ "suppress_tokens": null,
112
+ "task_specific_params": null,
113
+ "temperature": 1.0,
114
+ "tf_legacy_loss": false,
115
+ "tie_encoder_decoder": false,
116
+ "tie_word_embeddings": false,
117
+ "tokenizer_class": "QWenTokenizer",
118
+ "top_k": 50,
119
+ "top_p": 1.0,
120
+ "torch_dtype": null,
121
+ "torchscript": false,
122
+ "transformers_version": "4.32.0",
123
+ "typical_p": 1.0,
124
+ "use_bfloat16": false,
125
+ "use_cache": true,
126
+ "use_cache_kernel": false,
127
+ "use_cache_quantization": false,
128
+ "use_dynamic_ntk": true,
129
+ "use_flash_attn": false,
130
+ "use_logn_attn": true,
131
+ "vocab_size": 151936
132
+ },
133
+ "torch_dtype": "bfloat16",
134
+ "transformers_version": "4.32.0",
135
+ "whisper_config": {
136
+ "_name_or_path": "openai/whisper-large-v2",
137
+ "activation_dropout": 0.0,
138
+ "activation_function": "gelu",
139
+ "add_cross_attention": false,
140
+ "apply_spec_augment": false,
141
+ "architectures": [
142
+ "WhisperForConditionalGeneration"
143
+ ],
144
+ "attention_dropout": 0.0,
145
+ "bad_words_ids": null,
146
+ "begin_suppress_tokens": [
147
+ 220,
148
+ 50257
149
+ ],
150
+ "bos_token_id": 50257,
151
+ "chunk_size_feed_forward": 0,
152
+ "classifier_proj_size": 256,
153
+ "cross_attention_hidden_size": null,
154
+ "d_model": 1280,
155
+ "decoder_attention_heads": 20,
156
+ "decoder_ffn_dim": 5120,
157
+ "decoder_layerdrop": 0.0,
158
+ "decoder_layers": 32,
159
+ "decoder_start_token_id": 50258,
160
+ "diversity_penalty": 0.0,
161
+ "do_sample": false,
162
+ "dropout": 0.0,
163
+ "early_stopping": false,
164
+ "encoder_attention_heads": 20,
165
+ "encoder_ffn_dim": 5120,
166
+ "encoder_layerdrop": 0.0,
167
+ "encoder_layers": 32,
168
+ "encoder_no_repeat_ngram_size": 0,
169
+ "eos_token_id": 50257,
170
+ "exponential_decay_length_penalty": null,
171
+ "finetuning_task": null,
172
+ "forced_bos_token_id": null,
173
+ "forced_decoder_ids": [
174
+ [
175
+ 1,
176
+ 50259
177
+ ],
178
+ [
179
+ 2,
180
+ 50359
181
+ ],
182
+ [
183
+ 3,
184
+ 50363
185
+ ]
186
+ ],
187
+ "forced_eos_token_id": null,
188
+ "id2label": {
189
+ "0": "LABEL_0",
190
+ "1": "LABEL_1"
191
+ },
192
+ "init_std": 0.02,
193
+ "is_decoder": false,
194
+ "is_encoder_decoder": true,
195
+ "label2id": {
196
+ "LABEL_0": 0,
197
+ "LABEL_1": 1
198
+ },
199
+ "length_penalty": 1.0,
200
+ "mask_feature_length": 10,
201
+ "mask_feature_min_masks": 0,
202
+ "mask_feature_prob": 0.0,
203
+ "mask_time_length": 10,
204
+ "mask_time_min_masks": 2,
205
+ "mask_time_prob": 0.05,
206
+ "max_length": 448,
207
+ "max_source_positions": 1500,
208
+ "max_target_positions": 448,
209
+ "median_filter_width": 7,
210
+ "min_length": 0,
211
+ "model_type": "whisper",
212
+ "no_repeat_ngram_size": 0,
213
+ "num_beam_groups": 1,
214
+ "num_beams": 1,
215
+ "num_hidden_layers": 32,
216
+ "num_mel_bins": 80,
217
+ "num_return_sequences": 1,
218
+ "output_attentions": false,
219
+ "output_hidden_states": false,
220
+ "output_scores": false,
221
+ "pad_token_id": 50257,
222
+ "prefix": null,
223
+ "problem_type": null,
224
+ "pruned_heads": {},
225
+ "remove_invalid_values": false,
226
+ "repetition_penalty": 1.0,
227
+ "return_dict": true,
228
+ "return_dict_in_generate": false,
229
+ "scale_embedding": false,
230
+ "sep_token_id": null,
231
+ "suppress_tokens": [
232
+ 1,
233
+ 2,
234
+ 7,
235
+ 8,
236
+ 9,
237
+ 10,
238
+ 14,
239
+ 25,
240
+ 26,
241
+ 27,
242
+ 28,
243
+ 29,
244
+ 31,
245
+ 58,
246
+ 59,
247
+ 60,
248
+ 61,
249
+ 62,
250
+ 63,
251
+ 90,
252
+ 91,
253
+ 92,
254
+ 93,
255
+ 359,
256
+ 503,
257
+ 522,
258
+ 542,
259
+ 873,
260
+ 893,
261
+ 902,
262
+ 918,
263
+ 922,
264
+ 931,
265
+ 1350,
266
+ 1853,
267
+ 1982,
268
+ 2460,
269
+ 2627,
270
+ 3246,
271
+ 3253,
272
+ 3268,
273
+ 3536,
274
+ 3846,
275
+ 3961,
276
+ 4183,
277
+ 4667,
278
+ 6585,
279
+ 6647,
280
+ 7273,
281
+ 9061,
282
+ 9383,
283
+ 10428,
284
+ 10929,
285
+ 11938,
286
+ 12033,
287
+ 12331,
288
+ 12562,
289
+ 13793,
290
+ 14157,
291
+ 14635,
292
+ 15265,
293
+ 15618,
294
+ 16553,
295
+ 16604,
296
+ 18362,
297
+ 18956,
298
+ 20075,
299
+ 21675,
300
+ 22520,
301
+ 26130,
302
+ 26161,
303
+ 26435,
304
+ 28279,
305
+ 29464,
306
+ 31650,
307
+ 32302,
308
+ 32470,
309
+ 36865,
310
+ 42863,
311
+ 47425,
312
+ 49870,
313
+ 50254,
314
+ 50258,
315
+ 50358,
316
+ 50359,
317
+ 50360,
318
+ 50361,
319
+ 50362
320
+ ],
321
+ "task_specific_params": null,
322
+ "temperature": 1.0,
323
+ "tf_legacy_loss": false,
324
+ "tie_encoder_decoder": false,
325
+ "tie_word_embeddings": true,
326
+ "tokenizer_class": null,
327
+ "top_k": 50,
328
+ "top_p": 1.0,
329
+ "torch_dtype": "float32",
330
+ "torchscript": false,
331
+ "transformers_version": "4.32.0",
332
+ "typical_p": 1.0,
333
+ "use_bfloat16": false,
334
+ "use_cache": true,
335
+ "use_weighted_layer_sum": false,
336
+ "vocab_size": 51865
337
+ }
338
+ }
generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "chat_format": "chatml",
3
+ "do_sample": true,
4
+ "eos_token_id": 151643,
5
+ "max_new_tokens": 512,
6
+ "max_window_size": 6144,
7
+ "pad_token_id": 151643,
8
+ "top_k": 0,
9
+ "top_p": 0.5,
10
+ "transformers_version": "4.32.0"
11
+ }
preprocessor_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "chunk_length": 30,
3
+ "feature_extractor_type": "WhisperFeatureExtractor",
4
+ "feature_size": 80,
5
+ "hop_length": 160,
6
+ "n_fft": 400,
7
+ "n_samples": 480000,
8
+ "nb_max_frames": 3000,
9
+ "padding_side": "right",
10
+ "padding_value": 0.0,
11
+ "processor_class": "WhisperProcessor",
12
+ "return_attention_mask": false,
13
+ "sampling_rate": 16000
14
+ }
pytorch_model.bin.index.json ADDED
The diff for this file is too large to render. See raw diff
 
qwen.tiktoken ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {}
tokenizer_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_map": {
3
+ "AutoTokenizer": [
4
+ "tokenization_qwen.QWenTokenizer",
5
+ null
6
+ ]
7
+ },
8
+ "clean_up_tokenization_spaces": true,
9
+ "model_max_length": 8192,
10
+ "tokenizer_class": "QWenTokenizer"
11
+ }