wenhuach commited on
Commit
07a117c
1 Parent(s): 327644c

replace with autogptq format

Browse files

Signed-off-by: wenhuach <wenhuach87@gmail.com>

Files changed (3) hide show
  1. config.json +12 -9
  2. generation_config.json +1 -1
  3. model.safetensors +2 -2
config.json CHANGED
@@ -18,12 +18,13 @@
18
  "num_key_value_heads": 4,
19
  "quantization_config": {
20
  "amp": true,
21
- "autoround_version": "0.2.1.dev",
22
- "backend": "autoround:exllamav2",
23
  "bits": 4,
 
24
  "data_type": "int",
25
- "dataset": "NeelNanda/pile-10k",
26
  "enable_minmax_tuning": true,
 
27
  "enable_quanted_input": true,
28
  "gradient_accumulate_steps": 1,
29
  "group_size": 128,
@@ -31,19 +32,21 @@
31
  "low_gpu_mem_usage": false,
32
  "lr": 0.001,
33
  "minmax_lr": 0.001,
34
- "n_samples": 512,
35
- "quant_method": "intel/auto-round",
 
36
  "scale_dtype": "torch.float16",
37
  "seqlen": 2048,
38
- "sym": false,
39
- "train_bs": 8
 
40
  },
41
  "rms_norm_eps": 1e-06,
42
  "rope_theta": 1000000.0,
43
- "sliding_window": 131072,
44
  "tie_word_embeddings": false,
45
  "torch_dtype": "float16",
46
- "transformers_version": "4.40.0",
47
  "use_cache": true,
48
  "use_sliding_window": false,
49
  "vocab_size": 152064
 
18
  "num_key_value_heads": 4,
19
  "quantization_config": {
20
  "amp": true,
21
+ "autoround_version": "0.3.1.dev",
 
22
  "bits": 4,
23
+ "damp_percent": 0.01,
24
  "data_type": "int",
25
+ "desc_act": false,
26
  "enable_minmax_tuning": true,
27
+ "enable_norm_bias_tuning": false,
28
  "enable_quanted_input": true,
29
  "gradient_accumulate_steps": 1,
30
  "group_size": 128,
 
32
  "low_gpu_mem_usage": false,
33
  "lr": 0.001,
34
  "minmax_lr": 0.001,
35
+ "nsamples": 512,
36
+ "quant_block_list": null,
37
+ "quant_method": "gptq",
38
  "scale_dtype": "torch.float16",
39
  "seqlen": 2048,
40
+ "sym": true,
41
+ "train_bs": 8,
42
+ "true_sequential": false
43
  },
44
  "rms_norm_eps": 1e-06,
45
  "rope_theta": 1000000.0,
46
+ "sliding_window": null,
47
  "tie_word_embeddings": false,
48
  "torch_dtype": "float16",
49
+ "transformers_version": "4.44.2",
50
  "use_cache": true,
51
  "use_sliding_window": false,
52
  "vocab_size": 152064
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "bos_token_id": 151643,
3
  "eos_token_id": 151643,
4
  "max_new_tokens": 2048,
5
- "transformers_version": "4.40.0"
6
  }
 
2
  "bos_token_id": 151643,
3
  "eos_token_id": 151643,
4
  "max_new_tokens": 2048,
5
+ "transformers_version": "4.44.2"
6
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a8859aa3ec2bd85f859fe7cf264fafa2a75f912ae9f6e1fa8a739d490dc64345
3
- size 5575381304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e862995baefeb5e7281c07fa1011889b579e73bd2d9804faa30034845c1ac70
3
+ size 5573365008