Xu-Ouyang commited on
Commit
e7830ff
1 Parent(s): d780515

Upload GPTNeoXForCausalLM

Browse files
Files changed (3) hide show
  1. config.json +12 -3
  2. generation_config.json +1 -1
  3. model.safetensors +1 -1
config.json CHANGED
@@ -19,22 +19,31 @@
19
  "num_attention_heads": 8,
20
  "num_hidden_layers": 16,
21
  "quantization_config": {
 
22
  "bits": 8,
 
23
  "damp_percent": 0.1,
24
  "dataset": "wikitext2",
25
  "desc_act": false,
 
26
  "group_size": 128,
27
- "modules_in_block_to_quantize": null,
 
 
 
 
 
28
  "quant_method": "gptq",
29
  "sym": true,
30
- "true_sequential": true
 
31
  },
32
  "rope_scaling": null,
33
  "rotary_emb_base": 10000,
34
  "rotary_pct": 0.25,
35
  "tie_word_embeddings": false,
36
  "torch_dtype": "float16",
37
- "transformers_version": "4.41.2",
38
  "use_cache": true,
39
  "use_parallel_residual": true,
40
  "vocab_size": 50304
 
19
  "num_attention_heads": 8,
20
  "num_hidden_layers": 16,
21
  "quantization_config": {
22
+ "batch_size": 1,
23
  "bits": 8,
24
+ "block_name_to_quantize": "gpt_neox.layers",
25
  "damp_percent": 0.1,
26
  "dataset": "wikitext2",
27
  "desc_act": false,
28
+ "disable_exllama": false,
29
  "group_size": 128,
30
+ "model_seqlen": 2048,
31
+ "module_name_preceding_first_block": [
32
+ "gpt_neox.embed_in",
33
+ "gpt_neox.emb_dropout"
34
+ ],
35
+ "pad_token_id": null,
36
  "quant_method": "gptq",
37
  "sym": true,
38
+ "true_sequential": true,
39
+ "use_cuda_fp16": true
40
  },
41
  "rope_scaling": null,
42
  "rotary_emb_base": 10000,
43
  "rotary_pct": 0.25,
44
  "tie_word_embeddings": false,
45
  "torch_dtype": "float16",
46
+ "transformers_version": "4.44.0",
47
  "use_cache": true,
48
  "use_parallel_residual": true,
49
  "vocab_size": 50304
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
  "eos_token_id": 0,
5
- "transformers_version": "4.41.2"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
  "eos_token_id": 0,
5
+ "transformers_version": "4.44.0"
6
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c602b04807a21941395e0b7fd2d08e77cca51e039fa3373f9c8f79a88c34fc74
3
  size 1238094144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae026ad8156556deae200004ed2cb946ee057423c9b3e7ec60159672fb3d5791
3
  size 1238094144