Upload folder using huggingface_hub (#1)
Browse files- 9f7ccf3a1b5869265c440ec7687abd8d4a79f092ee64ab15aa01a93587adaea3 (e0e7a350211d18cf4a18a84d00cbd6a15758fc27)
- 8eb74b05f4beffb0d08096ab9d32313593057ac87257ab48c23575be8a67f59a (8c976293d155187de71580b82d88466031e8c684)
- f357b5b3381eb49acb143524eccaad08b302160e70663db26ceacc1447f5fdcb (69f1e702e739de7f37db8edb6c72fc72ad0c27d2)
- 7b49a823aa544e8a754ab4bc77a21d29575a7a0582dfefda57c2ae697fcff458 (c44b517bbebd7d4d5d81b957e7bdca1a086e8a40)
- 5291561befb4e96de04dced2cf371826f7bb41c31c1bafc9a46482e4db41d04e (c5f468fe823fe677aa9c5ec37eabdd7e9c206906)
- 8935840981dffc898ed5bb26865ae8aef832ee96671da0ad92f809553575cf9f (951daed1453f6bdac98884e6a18e10adc51d07e2)
- 45dcb1560f693f5b4d58cae76b17b8d20c672151b1f083d3fafdb9541a2740d1 (1a4761eccfeb7ff1a145c46ac5840af913a4eb3e)
- be2e19a6042958e9879b92a2ddff25d6b3b6ef43566faa1b2fa7417909b6b7e4 (9354e04564a88b19784544e6431f58b8023d9ac4)
- c66f9530135f5dc5465720b589031610d27f9b500a8c22d62f99c1283d27a641 (fa6c2aa5964ec25a336b8b6c3c652d89df6981cf)
- a6d69bbd712d784f0063151d955e2096bc20695c039d325f7562511487fc9d66 (a9c4ebddd2667818b2a076a71f38b2ffc2a9a540)
- d5959d2606681adbe7dbca5563c8b791b7feade89947e6e4a02d7890379884dd (8117ecf94568cd6f3f7d5513735630e1883635e4)
- 17a871e0045ed633579d237c511850616bd219d258a9a5c73fe7bf5fe68b63c2 (c4aa0480f436581edb98267ae18850224c03866a)
- c8de80f973728ade8456ff9bfaa3ff929fce085910ab3e62d4e5f61ce17ded74 (53716b44fde61ad168781e0ecee52c89eda02847)
- b7dd69b4a4184ad0d7ca8b8cfb8d3463b68c927f7b5af322833f60736cf9b04f (907af4513d568036390cef237d3144752f1052d5)
- e838c653f46e6b7de60f161f3027d22ba5f4092c90a4d688dfbb605e3df26880 (3b319e46b9c966b1cad69c87de3a5da5dca0b039)
- 4ef345d6a1cdbcbaf363095fe970794c2477d6a6d4414d79c205c42edc226d18 (b16ad286c6ff082a6b7e43cd1cc3fab1b4f608b5)
- README.md +32 -0
- config.json +38 -0
- model-00001-of-00016.safetensors +3 -0
- model-00002-of-00016.safetensors +3 -0
- model-00003-of-00016.safetensors +3 -0
- model-00004-of-00016.safetensors +3 -0
- model-00005-of-00016.safetensors +3 -0
- model-00006-of-00016.safetensors +3 -0
- model-00007-of-00016.safetensors +3 -0
- model-00008-of-00016.safetensors +3 -0
- model-00009-of-00016.safetensors +3 -0
- model-00010-of-00016.safetensors +3 -0
- model-00011-of-00016.safetensors +3 -0
- model-00012-of-00016.safetensors +3 -0
- model-00013-of-00016.safetensors +3 -0
- model-00014-of-00016.safetensors +3 -0
- model-00015-of-00016.safetensors +3 -0
- model-00016-of-00016.safetensors +3 -0
- model.safetensors.index.json +0 -0
- special_tokens_map.json +30 -0
- tokenizer.json +0 -0
- tokenizer.model +3 -0
- tokenizer_config.json +45 -0
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model: rAIfle/SorcererLM-8x22b-bf16
|
3 |
+
license: apache-2.0
|
4 |
+
tags:
|
5 |
+
- mlx
|
6 |
+
---
|
7 |
+
|
8 |
+
# mlx-community/SorcererLM-8x22b-4bit
|
9 |
+
|
10 |
+
The Model [mlx-community/SorcererLM-8x22b-4bit](https://huggingface.co/mlx-community/SorcererLM-8x22b-4bit) was converted to MLX format from [rAIfle/SorcererLM-8x22b-bf16](https://huggingface.co/rAIfle/SorcererLM-8x22b-bf16) using mlx-lm version **0.19.0**.
|
11 |
+
|
12 |
+
## Use with mlx
|
13 |
+
|
14 |
+
```bash
|
15 |
+
pip install mlx-lm
|
16 |
+
```
|
17 |
+
|
18 |
+
```python
|
19 |
+
from mlx_lm import load, generate
|
20 |
+
|
21 |
+
model, tokenizer = load("mlx-community/SorcererLM-8x22b-4bit")
|
22 |
+
|
23 |
+
prompt="hello"
|
24 |
+
|
25 |
+
if hasattr(tokenizer, "apply_chat_template") and tokenizer.chat_template is not None:
|
26 |
+
messages = [{"role": "user", "content": prompt}]
|
27 |
+
prompt = tokenizer.apply_chat_template(
|
28 |
+
messages, tokenize=False, add_generation_prompt=True
|
29 |
+
)
|
30 |
+
|
31 |
+
response = generate(model, tokenizer, prompt=prompt, verbose=True)
|
32 |
+
```
|
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"MixtralForCausalLM"
|
4 |
+
],
|
5 |
+
"attention_dropout": 0.0,
|
6 |
+
"bos_token_id": 1,
|
7 |
+
"eos_token_id": 2,
|
8 |
+
"hidden_act": "silu",
|
9 |
+
"hidden_size": 6144,
|
10 |
+
"initializer_range": 0.02,
|
11 |
+
"intermediate_size": 16384,
|
12 |
+
"max_position_embeddings": 65536,
|
13 |
+
"model_type": "mixtral",
|
14 |
+
"num_attention_heads": 48,
|
15 |
+
"num_experts_per_tok": 2,
|
16 |
+
"num_hidden_layers": 56,
|
17 |
+
"num_key_value_heads": 8,
|
18 |
+
"num_local_experts": 8,
|
19 |
+
"output_router_logits": false,
|
20 |
+
"quantization": {
|
21 |
+
"group_size": 64,
|
22 |
+
"bits": 4
|
23 |
+
},
|
24 |
+
"quantization_config": {
|
25 |
+
"group_size": 64,
|
26 |
+
"bits": 4
|
27 |
+
},
|
28 |
+
"rms_norm_eps": 1e-05,
|
29 |
+
"rope_theta": 1000000,
|
30 |
+
"router_aux_loss_coef": 0.001,
|
31 |
+
"router_jitter_noise": 0.0,
|
32 |
+
"sliding_window": null,
|
33 |
+
"tie_word_embeddings": false,
|
34 |
+
"torch_dtype": "bfloat16",
|
35 |
+
"transformers_version": "4.45.0.dev0",
|
36 |
+
"use_cache": false,
|
37 |
+
"vocab_size": 32000
|
38 |
+
}
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7b46ac1e31220fa1cb133baf77df7da2976c9e94b9842f59d3eb95b1f078b507
|
3 |
+
size 5291802737
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7a2bc2a2d32a46a8d081b56657656273256d20c6efeca250f44df29a0001c754
|
3 |
+
size 5181235243
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e62237fd299e53cadfdbb82cd89978e95e483955c9103474de8437d5f94927b8
|
3 |
+
size 5181235208
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cf11f618c4e58e51d57b2cc441fe311d6e8ab76e795dc47b117a18efbdb18143
|
3 |
+
size 5131635931
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:03490b11a2822f2825e4238673d9690f0fe1a84d20e66e07e27df2fbe27c6a44
|
3 |
+
size 5181235326
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1ac453cb7ba06fb5efa478b2a9618f3996c8affd90a5aff4417f7e723a3a22bf
|
3 |
+
size 5181235312
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5b782d7398d10bb14e3f9c02bc7d231e625a707794316f9616244b9d063db00e
|
3 |
+
size 5131635949
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7a62120931c0c263b2bd92277985c4a0c076ecd7f0506f4f139dd8387f640fa3
|
3 |
+
size 5181235342
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c12053f53d0515fd5f24deed88f4127d3a1a9bb966bd54a3c683e09257e4f911
|
3 |
+
size 5181235294
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4bb89d969771943cb39e9a0a308260db3e0877365c5b65a104f59c7dfeabe0f2
|
3 |
+
size 5131635917
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba5a78dbdaa8a6293868ab5869ae2eca7953463121fb09b929bae6a54d174c0e
|
3 |
+
size 5181235346
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9798d8a3b8337e63378f7481d2bd32937cda52b35d65755cb7a034cc61d6f0d2
|
3 |
+
size 5181235292
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f50cb1d11d02a9f20468aa15e1fe2ca8ad203371ad74b217d8c6b754bb44fa05
|
3 |
+
size 5131635939
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6cef94b43129ef943a9c6de2139bef83a04b7b0cd3f67843001d368d17409364
|
3 |
+
size 5181235376
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:41040d08d596c9c24409f08e5e862d5be5230413d4d2dd0f3bf21f1ccfe274b5
|
3 |
+
size 5181235324
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6bb8239c951f4e8b1194ef69068036200b2c1d3fe16306de8d47171ce36b8741
|
3 |
+
size 1469585211
|
The diff for this file is too large to render.
See raw diff
|
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "</s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "<unk>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"unk_token": {
|
24 |
+
"content": "<unk>",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
}
|
30 |
+
}
|
The diff for this file is too large to render.
See raw diff
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
|
3 |
+
size 493443
|
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": true,
|
3 |
+
"add_eos_token": false,
|
4 |
+
"add_prefix_space": true,
|
5 |
+
"added_tokens_decoder": {
|
6 |
+
"0": {
|
7 |
+
"content": "<unk>",
|
8 |
+
"lstrip": false,
|
9 |
+
"normalized": false,
|
10 |
+
"rstrip": false,
|
11 |
+
"single_word": false,
|
12 |
+
"special": true
|
13 |
+
},
|
14 |
+
"1": {
|
15 |
+
"content": "<s>",
|
16 |
+
"lstrip": false,
|
17 |
+
"normalized": false,
|
18 |
+
"rstrip": false,
|
19 |
+
"single_word": false,
|
20 |
+
"special": true
|
21 |
+
},
|
22 |
+
"2": {
|
23 |
+
"content": "</s>",
|
24 |
+
"lstrip": false,
|
25 |
+
"normalized": false,
|
26 |
+
"rstrip": false,
|
27 |
+
"single_word": false,
|
28 |
+
"special": true
|
29 |
+
}
|
30 |
+
},
|
31 |
+
"additional_special_tokens": [],
|
32 |
+
"bos_token": "<s>",
|
33 |
+
"chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{{ messages[0]['content'].strip() }}{% else %}{% set loop_messages = messages %}{{ 'A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user\\'s questions.' }}{% endif %}{% for message in loop_messages %}{% if loop.index0 == 0 %}{% if message['role'] == 'system' or message['role'] == 'user' %}{{ ' USER: ' + message['content'].strip() }}{% else %}{{ ' ASSISTANT: ' + message['content'].strip() + eos_token }}{% endif %}{% else %}{% if message['role'] == 'system' or message['role'] == 'user' %}{{ '\nUSER: ' + message['content'].strip() }}{% else %}{{ ' ASSISTANT: ' + message['content'].strip() + eos_token }}{% endif %}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ ' ASSISTANT:' }}{% endif %}",
|
34 |
+
"clean_up_tokenization_spaces": false,
|
35 |
+
"eos_token": "</s>",
|
36 |
+
"legacy": true,
|
37 |
+
"model_max_length": 1000000000000000019884624838656,
|
38 |
+
"pad_token": "<unk>",
|
39 |
+
"padding_side": "right",
|
40 |
+
"sp_model_kwargs": {},
|
41 |
+
"spaces_between_special_tokens": false,
|
42 |
+
"tokenizer_class": "LlamaTokenizer",
|
43 |
+
"unk_token": "<unk>",
|
44 |
+
"use_default_system_prompt": true
|
45 |
+
}
|