repo init
Browse files- added_tokens.json +5 -0
- merges.txt +0 -0
- mlc-chat-config.json +78 -0
- ndarray-cache.json +0 -0
- params_shard_0.bin +3 -0
- params_shard_1.bin +3 -0
- params_shard_10.bin +3 -0
- params_shard_11.bin +3 -0
- params_shard_12.bin +3 -0
- params_shard_13.bin +3 -0
- params_shard_14.bin +3 -0
- params_shard_15.bin +3 -0
- params_shard_16.bin +3 -0
- params_shard_17.bin +3 -0
- params_shard_18.bin +3 -0
- params_shard_19.bin +3 -0
- params_shard_2.bin +3 -0
- params_shard_20.bin +3 -0
- params_shard_21.bin +3 -0
- params_shard_22.bin +3 -0
- params_shard_23.bin +3 -0
- params_shard_24.bin +3 -0
- params_shard_25.bin +3 -0
- params_shard_26.bin +3 -0
- params_shard_27.bin +3 -0
- params_shard_28.bin +3 -0
- params_shard_29.bin +3 -0
- params_shard_3.bin +3 -0
- params_shard_4.bin +3 -0
- params_shard_5.bin +3 -0
- params_shard_6.bin +3 -0
- params_shard_7.bin +3 -0
- params_shard_8.bin +3 -0
- params_shard_9.bin +3 -0
- tokenizer.json +0 -0
- tokenizer_config.json +43 -0
- vocab.json +0 -0
added_tokens.json
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"<|endoftext|>": 151643,
|
3 |
+
"<|im_end|>": 151645,
|
4 |
+
"<|im_start|>": 151644
|
5 |
+
}
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
mlc-chat-config.json
ADDED
@@ -0,0 +1,78 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"model_type": "qwen2",
|
3 |
+
"quantization": "q4f16_1",
|
4 |
+
"model_config": {
|
5 |
+
"hidden_act": "silu",
|
6 |
+
"hidden_size": 1536,
|
7 |
+
"intermediate_size": 8960,
|
8 |
+
"num_attention_heads": 12,
|
9 |
+
"num_hidden_layers": 28,
|
10 |
+
"num_key_value_heads": 2,
|
11 |
+
"rms_norm_eps": 1e-06,
|
12 |
+
"rope_theta": 1000000.0,
|
13 |
+
"vocab_size": 151936,
|
14 |
+
"context_window_size": 768,
|
15 |
+
"prefill_chunk_size": 768,
|
16 |
+
"tensor_parallel_shards": 1,
|
17 |
+
"head_dim": 128,
|
18 |
+
"dtype": "float32"
|
19 |
+
},
|
20 |
+
"vocab_size": 151936,
|
21 |
+
"context_window_size": 768,
|
22 |
+
"sliding_window_size": -1,
|
23 |
+
"prefill_chunk_size": 768,
|
24 |
+
"attention_sink_size": -1,
|
25 |
+
"tensor_parallel_shards": 1,
|
26 |
+
"mean_gen_len": 128,
|
27 |
+
"max_gen_len": 512,
|
28 |
+
"shift_fill_factor": 0.3,
|
29 |
+
"temperature": 0.7,
|
30 |
+
"presence_penalty": 0.0,
|
31 |
+
"frequency_penalty": 0.0,
|
32 |
+
"repetition_penalty": 1.1,
|
33 |
+
"top_p": 0.8,
|
34 |
+
"conv_template": {
|
35 |
+
"name": "chatml",
|
36 |
+
"system_template": "<|im_start|>system\n{system_message}",
|
37 |
+
"system_message": "A conversation between a user and an LLM-based AI assistant. The assistant gives helpful and honest answers.",
|
38 |
+
"add_role_after_system_message": true,
|
39 |
+
"roles": {
|
40 |
+
"user": "<|im_start|>user",
|
41 |
+
"assistant": "<|im_start|>assistant"
|
42 |
+
},
|
43 |
+
"role_templates": {
|
44 |
+
"user": "{user_message}",
|
45 |
+
"assistant": "{assistant_message}",
|
46 |
+
"tool": "{tool_message}"
|
47 |
+
},
|
48 |
+
"messages": [],
|
49 |
+
"seps": [
|
50 |
+
"<|im_end|>\n"
|
51 |
+
],
|
52 |
+
"role_content_sep": "\n",
|
53 |
+
"role_empty_sep": "\n",
|
54 |
+
"stop_str": [
|
55 |
+
"<|im_end|>"
|
56 |
+
],
|
57 |
+
"stop_token_ids": [
|
58 |
+
2
|
59 |
+
],
|
60 |
+
"function_string": "",
|
61 |
+
"use_function_calling": false,
|
62 |
+
"image_token_index": -1
|
63 |
+
},
|
64 |
+
"pad_token_id": 151643,
|
65 |
+
"bos_token_id": 151643,
|
66 |
+
"eos_token_id": [
|
67 |
+
151645,
|
68 |
+
151643
|
69 |
+
],
|
70 |
+
"tokenizer_files": [
|
71 |
+
"tokenizer.json",
|
72 |
+
"vocab.json",
|
73 |
+
"merges.txt",
|
74 |
+
"added_tokens.json",
|
75 |
+
"tokenizer_config.json"
|
76 |
+
],
|
77 |
+
"version": "0.1.0"
|
78 |
+
}
|
ndarray-cache.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
params_shard_0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a59e9857135c94004fee8d3003a7d55e8604bcdb5aaf38c862a2e833fb47306c
|
3 |
+
size 116686848
|
params_shard_1.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a59e9857135c94004fee8d3003a7d55e8604bcdb5aaf38c862a2e833fb47306c
|
3 |
+
size 116686848
|
params_shard_10.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ed74f0802eda9e480f70e2314d6830cdb42bf932ae790f2e8c64935bc658a2b7
|
3 |
+
size 26331136
|
params_shard_11.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3af07a0b51b23bf8897d7d9f9aa1847a5cf4803957b963a20d4c6ffb61c67479
|
3 |
+
size 26331136
|
params_shard_12.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:59bc00e7e4c24c7eb422a47b122491119c93efff2f89cee6766500bdec274811
|
3 |
+
size 26331136
|
params_shard_13.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ccdeaff3423f6ea0f0f172323d02b1c46d7b3efe9151ca944bf16bc6520a2c43
|
3 |
+
size 26331136
|
params_shard_14.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1fc798908ce5d790808ee5bbc6a6c55295ad2cc0566e785a2fb7ceaa039733e4
|
3 |
+
size 26331136
|
params_shard_15.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f7733e32ce1c21e388e223c0233921b2314f0f3d54097fa624363606c7fb0e8f
|
3 |
+
size 26331136
|
params_shard_16.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:342a61e8db52c4be2aa6dbb6b6f3f6e257c6f61d8609d39d3ffad7ef9778664b
|
3 |
+
size 32349184
|
params_shard_17.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a318c1a648beae43f3424cf62db410563b65f1f7fca1b458e30abde635cd29d3
|
3 |
+
size 31155200
|
params_shard_18.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:750542be451278278b5bc0c59f79b23216aa4b619c21beda53cd00455db28d37
|
3 |
+
size 33212416
|
params_shard_19.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ee9ff34c7c5cb6e91bc7bbeffeee6bae9ed49424bc77611bd3bc3f09c7f0abb7
|
3 |
+
size 27191296
|
params_shard_2.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:11bc0d4be6d27df8cda25748fe0a603434821861773bba31e4dd4773bd17ba49
|
3 |
+
size 29174784
|
params_shard_20.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:aa4c200a81475a6db414562dcf03b8c0d7444329204587f49398e15b82b577c1
|
3 |
+
size 26331136
|
params_shard_21.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f50ad297ecb1b00ad6f5560730953fa3ce7c7bd3aa76f356092f0684d751b0e3
|
3 |
+
size 26331136
|
params_shard_22.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:beb334bc61a0b088f17e31803c31241f0bfbc24a65e00870ca63ebff24d7f444
|
3 |
+
size 26331136
|
params_shard_23.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:443b39b3d2a960035ce9843d93149c99e942c57609fc33e9160721d52dd3e1d4
|
3 |
+
size 33218560
|
params_shard_24.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:37025c71b9011fb44d6509607454a2faaf2eda61dcf44812c98ee03d5703e575
|
3 |
+
size 27191296
|
params_shard_25.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:66cf4fac18abf708cad1f99af96e1fa0481458215380389128f5d9343a88dc90
|
3 |
+
size 26331136
|
params_shard_26.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d7642ee14a0adfa74bc4c679121b7dd68ca394869738b6c572de122493540262
|
3 |
+
size 26331136
|
params_shard_27.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:13a0ffa165c764765668debad57251d0cfbe22be13538bc36824c33aca116745
|
3 |
+
size 26331136
|
params_shard_28.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a9a7380d48792c3a54658e6876ba76176c241f07950fba6ff85c55aa7d8289b2
|
3 |
+
size 26331136
|
params_shard_29.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d4db3cbb08157d8ca510790c55a0c241d7e6245e28132aef7d730c23cf902b8
|
3 |
+
size 18589696
|
params_shard_3.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:600412348edf94fc12ceae030198fee4eadbd40f1a125b495b1e24b7b5fa163b
|
3 |
+
size 33212416
|
params_shard_4.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:02319c6431c7cd6aae9030921a96cef6b548b434deca046fb4b78886ff126bc1
|
3 |
+
size 27191296
|
params_shard_5.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d3cc695d6a5a99c5ae1d07bdbfcae10425b7c11fbc2292708b7c8e9b03ef740
|
3 |
+
size 26331136
|
params_shard_6.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5432b7a0338f74352ea7e84e7e439ca77f476ab83d81bbac9da76d38a2418def
|
3 |
+
size 26331136
|
params_shard_7.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:642ae58228088ebbc2521c4b53cf0c2ae43bef27b641b92631b2c6438aaf92e2
|
3 |
+
size 26331136
|
params_shard_8.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:88c80fa53c282f275cfc2eb8467f31ad87c8703f74c22411ff475a9e621f4e59
|
3 |
+
size 26331136
|
params_shard_9.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c6e8a021b305ec1d73118c54deba798fca66f6efb497b3f0a18a968b8fcdafab
|
3 |
+
size 26331136
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"added_tokens_decoder": {
|
4 |
+
"151643": {
|
5 |
+
"content": "<|endoftext|>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": false,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false,
|
10 |
+
"special": true
|
11 |
+
},
|
12 |
+
"151644": {
|
13 |
+
"content": "<|im_start|>",
|
14 |
+
"lstrip": false,
|
15 |
+
"normalized": false,
|
16 |
+
"rstrip": false,
|
17 |
+
"single_word": false,
|
18 |
+
"special": true
|
19 |
+
},
|
20 |
+
"151645": {
|
21 |
+
"content": "<|im_end|>",
|
22 |
+
"lstrip": false,
|
23 |
+
"normalized": false,
|
24 |
+
"rstrip": false,
|
25 |
+
"single_word": false,
|
26 |
+
"special": true
|
27 |
+
}
|
28 |
+
},
|
29 |
+
"additional_special_tokens": [
|
30 |
+
"<|im_start|>",
|
31 |
+
"<|im_end|>"
|
32 |
+
],
|
33 |
+
"bos_token": null,
|
34 |
+
"chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
35 |
+
"clean_up_tokenization_spaces": false,
|
36 |
+
"eos_token": "<|im_end|>",
|
37 |
+
"errors": "replace",
|
38 |
+
"model_max_length": 32768,
|
39 |
+
"pad_token": "<|endoftext|>",
|
40 |
+
"split_special_tokens": false,
|
41 |
+
"tokenizer_class": "Qwen2Tokenizer",
|
42 |
+
"unk_token": null
|
43 |
+
}
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|