Upload folder using huggingface_hub
Browse files- config.json +9 -7
- model-00001-of-00007.safetensors +1 -1
- model-00002-of-00007.safetensors +1 -1
- model-00003-of-00007.safetensors +1 -1
- model-00004-of-00007.safetensors +1 -1
- model-00005-of-00007.safetensors +1 -1
- model-00006-of-00007.safetensors +1 -1
- model-00007-of-00007.safetensors +1 -1
- tokenizer_config.json +0 -1
config.json
CHANGED
@@ -40,13 +40,15 @@
|
|
40 |
"quantization_config": {
|
41 |
"_load_in_4bit": true,
|
42 |
"_load_in_8bit": false,
|
43 |
-
"bnb_4bit_compute_dtype": "
|
44 |
-
"bnb_4bit_quant_storage": "
|
45 |
-
"bnb_4bit_quant_type": "
|
46 |
-
"bnb_4bit_use_double_quant":
|
47 |
"llm_int8_enable_fp32_cpu_offload": false,
|
48 |
"llm_int8_has_fp16_weight": false,
|
49 |
-
"llm_int8_skip_modules":
|
|
|
|
|
50 |
"llm_int8_threshold": 6.0,
|
51 |
"load_in_4bit": true,
|
52 |
"load_in_8bit": false,
|
@@ -56,9 +58,9 @@
|
|
56 |
"router_aux_loss_coef": 0.001,
|
57 |
"sliding_window": null,
|
58 |
"tie_word_embeddings": false,
|
59 |
-
"torch_dtype": "
|
60 |
"transformers_version": "4.40.0.dev0",
|
61 |
-
"use_cache":
|
62 |
"use_mamba_kernels": true,
|
63 |
"vocab_size": 65536
|
64 |
}
|
|
|
40 |
"quantization_config": {
|
41 |
"_load_in_4bit": true,
|
42 |
"_load_in_8bit": false,
|
43 |
+
"bnb_4bit_compute_dtype": "float32",
|
44 |
+
"bnb_4bit_quant_storage": "uint8",
|
45 |
+
"bnb_4bit_quant_type": "fp4",
|
46 |
+
"bnb_4bit_use_double_quant": false,
|
47 |
"llm_int8_enable_fp32_cpu_offload": false,
|
48 |
"llm_int8_has_fp16_weight": false,
|
49 |
+
"llm_int8_skip_modules": [
|
50 |
+
"mamba"
|
51 |
+
],
|
52 |
"llm_int8_threshold": 6.0,
|
53 |
"load_in_4bit": true,
|
54 |
"load_in_8bit": false,
|
|
|
58 |
"router_aux_loss_coef": 0.001,
|
59 |
"sliding_window": null,
|
60 |
"tie_word_embeddings": false,
|
61 |
+
"torch_dtype": "float16",
|
62 |
"transformers_version": "4.40.0.dev0",
|
63 |
+
"use_cache": true,
|
64 |
"use_mamba_kernels": true,
|
65 |
"vocab_size": 65536
|
66 |
}
|
model-00001-of-00007.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4873003324
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2e92c801b498d391703c2d2aa7ecfd5251c43cb8fcea5469a35cba13e6e91fb8
|
3 |
size 4873003324
|
model-00002-of-00007.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4984623243
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b5931dc5bfcf2db3a287940e23861ffb9a9661e76e824cedb0a73d8cc359c2f0
|
3 |
size 4984623243
|
model-00003-of-00007.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4983413752
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ec106651959c1fc61482e954d6c1dee67d82d387f8c7edf366a1b89f9d14e027
|
3 |
size 4983413752
|
model-00004-of-00007.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4997975503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5f5178778f3ee8501126fe41ba37b8d7fd3837fcc5b2e902779ddba7e801ed42
|
3 |
size 4997975503
|
model-00005-of-00007.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4983413840
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ea765123548c90cc441e70a0c8f6d766208994730bf03f41c6b4f18be1f5ed9a
|
3 |
size 4983413840
|
model-00006-of-00007.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4975844741
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f25d10f17ca3965392e5c876a59d901d89a18bed479e8fa790c375ac3b40a9a3
|
3 |
size 4975844741
|
model-00007-of-00007.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3842901815
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:24d0e9a3fb0a6f8c61489778097d6debaf5da9ccc77299d1fef5fbb19b02ae9c
|
3 |
size 3842901815
|
tokenizer_config.json
CHANGED
@@ -36,7 +36,6 @@
|
|
36 |
}
|
37 |
},
|
38 |
"bos_token": "<|startoftext|>",
|
39 |
-
"chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
40 |
"clean_up_tokenization_spaces": false,
|
41 |
"eos_token": "<|endoftext|>",
|
42 |
"model_max_length": 1000000000000000019884624838656,
|
|
|
36 |
}
|
37 |
},
|
38 |
"bos_token": "<|startoftext|>",
|
|
|
39 |
"clean_up_tokenization_spaces": false,
|
40 |
"eos_token": "<|endoftext|>",
|
41 |
"model_max_length": 1000000000000000019884624838656,
|