morgul commited on
Commit
99e9534
1 Parent(s): ef8c8ec

Upload folder using huggingface_hub (#1)

Browse files

- 9f7ccf3a1b5869265c440ec7687abd8d4a79f092ee64ab15aa01a93587adaea3 (e0e7a350211d18cf4a18a84d00cbd6a15758fc27)
- 8eb74b05f4beffb0d08096ab9d32313593057ac87257ab48c23575be8a67f59a (8c976293d155187de71580b82d88466031e8c684)
- f357b5b3381eb49acb143524eccaad08b302160e70663db26ceacc1447f5fdcb (69f1e702e739de7f37db8edb6c72fc72ad0c27d2)
- 7b49a823aa544e8a754ab4bc77a21d29575a7a0582dfefda57c2ae697fcff458 (c44b517bbebd7d4d5d81b957e7bdca1a086e8a40)
- 5291561befb4e96de04dced2cf371826f7bb41c31c1bafc9a46482e4db41d04e (c5f468fe823fe677aa9c5ec37eabdd7e9c206906)
- 8935840981dffc898ed5bb26865ae8aef832ee96671da0ad92f809553575cf9f (951daed1453f6bdac98884e6a18e10adc51d07e2)
- 45dcb1560f693f5b4d58cae76b17b8d20c672151b1f083d3fafdb9541a2740d1 (1a4761eccfeb7ff1a145c46ac5840af913a4eb3e)
- be2e19a6042958e9879b92a2ddff25d6b3b6ef43566faa1b2fa7417909b6b7e4 (9354e04564a88b19784544e6431f58b8023d9ac4)
- c66f9530135f5dc5465720b589031610d27f9b500a8c22d62f99c1283d27a641 (fa6c2aa5964ec25a336b8b6c3c652d89df6981cf)
- a6d69bbd712d784f0063151d955e2096bc20695c039d325f7562511487fc9d66 (a9c4ebddd2667818b2a076a71f38b2ffc2a9a540)
- d5959d2606681adbe7dbca5563c8b791b7feade89947e6e4a02d7890379884dd (8117ecf94568cd6f3f7d5513735630e1883635e4)
- 17a871e0045ed633579d237c511850616bd219d258a9a5c73fe7bf5fe68b63c2 (c4aa0480f436581edb98267ae18850224c03866a)
- c8de80f973728ade8456ff9bfaa3ff929fce085910ab3e62d4e5f61ce17ded74 (53716b44fde61ad168781e0ecee52c89eda02847)
- b7dd69b4a4184ad0d7ca8b8cfb8d3463b68c927f7b5af322833f60736cf9b04f (907af4513d568036390cef237d3144752f1052d5)
- e838c653f46e6b7de60f161f3027d22ba5f4092c90a4d688dfbb605e3df26880 (3b319e46b9c966b1cad69c87de3a5da5dca0b039)
- 4ef345d6a1cdbcbaf363095fe970794c2477d6a6d4414d79c205c42edc226d18 (b16ad286c6ff082a6b7e43cd1cc3fab1b4f608b5)

README.md ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: rAIfle/SorcererLM-8x22b-bf16
3
+ license: apache-2.0
4
+ tags:
5
+ - mlx
6
+ ---
7
+
8
+ # mlx-community/SorcererLM-8x22b-4bit
9
+
10
+ The Model [mlx-community/SorcererLM-8x22b-4bit](https://huggingface.co/mlx-community/SorcererLM-8x22b-4bit) was converted to MLX format from [rAIfle/SorcererLM-8x22b-bf16](https://huggingface.co/rAIfle/SorcererLM-8x22b-bf16) using mlx-lm version **0.19.0**.
11
+
12
+ ## Use with mlx
13
+
14
+ ```bash
15
+ pip install mlx-lm
16
+ ```
17
+
18
+ ```python
19
+ from mlx_lm import load, generate
20
+
21
+ model, tokenizer = load("mlx-community/SorcererLM-8x22b-4bit")
22
+
23
+ prompt="hello"
24
+
25
+ if hasattr(tokenizer, "apply_chat_template") and tokenizer.chat_template is not None:
26
+ messages = [{"role": "user", "content": prompt}]
27
+ prompt = tokenizer.apply_chat_template(
28
+ messages, tokenize=False, add_generation_prompt=True
29
+ )
30
+
31
+ response = generate(model, tokenizer, prompt=prompt, verbose=True)
32
+ ```
config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MixtralForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 6144,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 16384,
12
+ "max_position_embeddings": 65536,
13
+ "model_type": "mixtral",
14
+ "num_attention_heads": 48,
15
+ "num_experts_per_tok": 2,
16
+ "num_hidden_layers": 56,
17
+ "num_key_value_heads": 8,
18
+ "num_local_experts": 8,
19
+ "output_router_logits": false,
20
+ "quantization": {
21
+ "group_size": 64,
22
+ "bits": 4
23
+ },
24
+ "quantization_config": {
25
+ "group_size": 64,
26
+ "bits": 4
27
+ },
28
+ "rms_norm_eps": 1e-05,
29
+ "rope_theta": 1000000,
30
+ "router_aux_loss_coef": 0.001,
31
+ "router_jitter_noise": 0.0,
32
+ "sliding_window": null,
33
+ "tie_word_embeddings": false,
34
+ "torch_dtype": "bfloat16",
35
+ "transformers_version": "4.45.0.dev0",
36
+ "use_cache": false,
37
+ "vocab_size": 32000
38
+ }
model-00001-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b46ac1e31220fa1cb133baf77df7da2976c9e94b9842f59d3eb95b1f078b507
3
+ size 5291802737
model-00002-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a2bc2a2d32a46a8d081b56657656273256d20c6efeca250f44df29a0001c754
3
+ size 5181235243
model-00003-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e62237fd299e53cadfdbb82cd89978e95e483955c9103474de8437d5f94927b8
3
+ size 5181235208
model-00004-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf11f618c4e58e51d57b2cc441fe311d6e8ab76e795dc47b117a18efbdb18143
3
+ size 5131635931
model-00005-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03490b11a2822f2825e4238673d9690f0fe1a84d20e66e07e27df2fbe27c6a44
3
+ size 5181235326
model-00006-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ac453cb7ba06fb5efa478b2a9618f3996c8affd90a5aff4417f7e723a3a22bf
3
+ size 5181235312
model-00007-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b782d7398d10bb14e3f9c02bc7d231e625a707794316f9616244b9d063db00e
3
+ size 5131635949
model-00008-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a62120931c0c263b2bd92277985c4a0c076ecd7f0506f4f139dd8387f640fa3
3
+ size 5181235342
model-00009-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c12053f53d0515fd5f24deed88f4127d3a1a9bb966bd54a3c683e09257e4f911
3
+ size 5181235294
model-00010-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bb89d969771943cb39e9a0a308260db3e0877365c5b65a104f59c7dfeabe0f2
3
+ size 5131635917
model-00011-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba5a78dbdaa8a6293868ab5869ae2eca7953463121fb09b929bae6a54d174c0e
3
+ size 5181235346
model-00012-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9798d8a3b8337e63378f7481d2bd32937cda52b35d65755cb7a034cc61d6f0d2
3
+ size 5181235292
model-00013-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f50cb1d11d02a9f20468aa15e1fe2ca8ad203371ad74b217d8c6b754bb44fa05
3
+ size 5131635939
model-00014-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cef94b43129ef943a9c6de2139bef83a04b7b0cd3f67843001d368d17409364
3
+ size 5181235376
model-00015-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41040d08d596c9c24409f08e5e862d5be5230413d4d2dd0f3bf21f1ccfe274b5
3
+ size 5181235324
model-00016-of-00016.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bb8239c951f4e8b1194ef69068036200b2c1d3fe16306de8d47171ce36b8741
3
+ size 1469585211
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
tokenizer_config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": true,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "additional_special_tokens": [],
32
+ "bos_token": "<s>",
33
+ "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{{ messages[0]['content'].strip() }}{% else %}{% set loop_messages = messages %}{{ 'A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user\\'s questions.' }}{% endif %}{% for message in loop_messages %}{% if loop.index0 == 0 %}{% if message['role'] == 'system' or message['role'] == 'user' %}{{ ' USER: ' + message['content'].strip() }}{% else %}{{ ' ASSISTANT: ' + message['content'].strip() + eos_token }}{% endif %}{% else %}{% if message['role'] == 'system' or message['role'] == 'user' %}{{ '\nUSER: ' + message['content'].strip() }}{% else %}{{ ' ASSISTANT: ' + message['content'].strip() + eos_token }}{% endif %}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ ' ASSISTANT:' }}{% endif %}",
34
+ "clean_up_tokenization_spaces": false,
35
+ "eos_token": "</s>",
36
+ "legacy": true,
37
+ "model_max_length": 1000000000000000019884624838656,
38
+ "pad_token": "<unk>",
39
+ "padding_side": "right",
40
+ "sp_model_kwargs": {},
41
+ "spaces_between_special_tokens": false,
42
+ "tokenizer_class": "LlamaTokenizer",
43
+ "unk_token": "<unk>",
44
+ "use_default_system_prompt": true
45
+ }