MLX
Safetensors
llama

[WIP] Upload folder using huggingface_hub (multi-commit 4b55d796901e10996dba682352f93059c36ffe1c87736f60096ac6f244deca54)

#1
README.md DELETED
@@ -1,24 +0,0 @@
1
- ---
2
- license: apache-2.0
3
- tags:
4
- - mlx
5
- ---
6
-
7
- # mlx-community/Yi-1.5-34B-4bit
8
-
9
- The Model [mlx-community/Yi-1.5-34B-4bit](https://huggingface.co/mlx-community/Yi-1.5-34B-4bit) was converted to MLX format from [01-ai/Yi-1.5-34B](https://huggingface.co/01-ai/Yi-1.5-34B) using mlx-lm version **0.13.0**.
10
-
11
- Model added by [Prince Canuma](https://twitter.com/Prince_Canuma).
12
-
13
- ## Use with mlx
14
-
15
- ```bash
16
- pip install mlx-lm
17
- ```
18
-
19
- ```python
20
- from mlx_lm import load, generate
21
-
22
- model, tokenizer = load("mlx-community/Yi-1.5-34B-4bit")
23
- response = generate(model, tokenizer, prompt="hello", verbose=True)
24
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
config.json DELETED
@@ -1,32 +0,0 @@
1
- {
2
- "architectures": [
3
- "LlamaForCausalLM"
4
- ],
5
- "attention_bias": false,
6
- "attention_dropout": 0.0,
7
- "bos_token_id": 1,
8
- "eos_token_id": 2,
9
- "hidden_act": "silu",
10
- "hidden_size": 7168,
11
- "initializer_range": 0.02,
12
- "intermediate_size": 20480,
13
- "max_position_embeddings": 4096,
14
- "model_type": "llama",
15
- "num_attention_heads": 56,
16
- "num_hidden_layers": 60,
17
- "num_key_value_heads": 8,
18
- "pad_token_id": 0,
19
- "pretraining_tp": 1,
20
- "quantization": {
21
- "group_size": 64,
22
- "bits": 4
23
- },
24
- "rms_norm_eps": 1e-06,
25
- "rope_scaling": null,
26
- "rope_theta": 5000000.0,
27
- "tie_word_embeddings": false,
28
- "torch_dtype": "bfloat16",
29
- "transformers_version": "4.37.2",
30
- "use_cache": true,
31
- "vocab_size": 64000
32
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
model-00001-of-00004.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9d8dfce543abd22bd8dcd404c8216607d0f121e1bb0257ca8a34205dace973e8
3
- size 5345192455
 
 
 
 
model-00002-of-00004.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:061a66727942baf467c1cc625a49bbe41257ae1c1d31d04c78c14032ab388a77
3
- size 5334900439
 
 
 
 
model-00003-of-00004.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:cfb92927305d6312e0d439c2c4c19a5ba8154cd0b67e419363af2ff426c5f3e3
3
- size 5334900447
 
 
 
 
model-00004-of-00004.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ba8a97ae7b148400b9b0f32a039699f92b83cc8ead96146338af2d9eafa6faf
3
- size 3330177746
 
 
 
 
model.safetensors.index.json DELETED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json DELETED
@@ -1,30 +0,0 @@
1
- {
2
- "bos_token": {
3
- "content": "<|startoftext|>",
4
- "lstrip": false,
5
- "normalized": false,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "<|endoftext|>",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "<unk>",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
- "unk_token": {
24
- "content": "<unk>",
25
- "lstrip": false,
26
- "normalized": false,
27
- "rstrip": false,
28
- "single_word": false
29
- }
30
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
tokenizer.model DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:386c49cf943d71aa110361135338c50e38beeff0a66593480421f37b319e1a39
3
- size 1033105
 
 
 
 
tokenizer_config.json DELETED
@@ -1,40 +0,0 @@
1
- {
2
- "add_bos_token": false,
3
- "add_eos_token": false,
4
- "added_tokens_decoder": {
5
- "0": {
6
- "content": "<unk>",
7
- "lstrip": false,
8
- "normalized": false,
9
- "rstrip": false,
10
- "single_word": false,
11
- "special": true
12
- },
13
- "1": {
14
- "content": "<|startoftext|>",
15
- "lstrip": false,
16
- "normalized": false,
17
- "rstrip": false,
18
- "single_word": false,
19
- "special": true
20
- },
21
- "2": {
22
- "content": "<|endoftext|>",
23
- "lstrip": false,
24
- "normalized": false,
25
- "rstrip": false,
26
- "single_word": false,
27
- "special": true
28
- }
29
- },
30
- "bos_token": "<|startoftext|>",
31
- "clean_up_tokenization_spaces": false,
32
- "eos_token": "<|endoftext|>",
33
- "legacy": true,
34
- "model_max_length": 4096,
35
- "pad_token": "<unk>",
36
- "sp_model_kwargs": {},
37
- "tokenizer_class": "LlamaTokenizer",
38
- "unk_token": "<unk>",
39
- "use_default_system_prompt": false
40
- }