Xu-Ouyang commited on
Commit
4359008
1 Parent(s): fe21666

Upload GPTNeoXForCausalLM

Browse files
config.json CHANGED
@@ -19,22 +19,31 @@
19
  "num_attention_heads": 32,
20
  "num_hidden_layers": 32,
21
  "quantization_config": {
 
22
  "bits": 8,
 
23
  "damp_percent": 0.1,
24
  "dataset": "wikitext2",
25
  "desc_act": false,
 
26
  "group_size": 128,
27
- "modules_in_block_to_quantize": null,
 
 
 
 
 
28
  "quant_method": "gptq",
29
  "sym": true,
30
- "true_sequential": true
 
31
  },
32
  "rope_scaling": null,
33
  "rotary_emb_base": 10000,
34
  "rotary_pct": 0.25,
35
  "tie_word_embeddings": false,
36
  "torch_dtype": "float16",
37
- "transformers_version": "4.41.2",
38
  "use_cache": true,
39
  "use_parallel_residual": true,
40
  "vocab_size": 50432
 
19
  "num_attention_heads": 32,
20
  "num_hidden_layers": 32,
21
  "quantization_config": {
22
+ "batch_size": 1,
23
  "bits": 8,
24
+ "block_name_to_quantize": "gpt_neox.layers",
25
  "damp_percent": 0.1,
26
  "dataset": "wikitext2",
27
  "desc_act": false,
28
+ "disable_exllama": false,
29
  "group_size": 128,
30
+ "model_seqlen": 4096,
31
+ "module_name_preceding_first_block": [
32
+ "gpt_neox.embed_in",
33
+ "gpt_neox.emb_dropout"
34
+ ],
35
+ "pad_token_id": null,
36
  "quant_method": "gptq",
37
  "sym": true,
38
+ "true_sequential": true,
39
+ "use_cuda_fp16": true
40
  },
41
  "rope_scaling": null,
42
  "rotary_emb_base": 10000,
43
  "rotary_pct": 0.25,
44
  "tie_word_embeddings": false,
45
  "torch_dtype": "float16",
46
+ "transformers_version": "4.44.0",
47
  "use_cache": true,
48
  "use_parallel_residual": true,
49
  "vocab_size": 50432
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
  "eos_token_id": 0,
5
- "transformers_version": "4.41.2"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
  "eos_token_id": 0,
5
+ "transformers_version": "4.44.0"
6
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48e4c1424e837b77b2accddec4b321099f5b696ca839b3372db017bfcc801dba
3
  size 4968290656
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca829ec36920a6a111303fd888d39d508c95d5aa70b43a47b2e1c9c55feced84
3
  size 4968290656
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a349c40affb5f7d3144ed838e6f2e358edd2a2a1c4bb24b50f8982df3909bc3a
3
  size 2458619040
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e00f01bbf9360b2f5b963288597ab68b4b6945a0e430f7b6b0e0a7d94b72e0d
3
  size 2458619040