cl-modelcloud commited on
Commit
d5526c2
1 Parent(s): 32ed296

Upload folder using huggingface_hub (#1)

Browse files

- 1404a834ae25c5958a5475b83ab98ae33d15cfd3ef34c31d26e99759d7c94986 (b0f05c6d91bcd0466412b9ba0bf9f70ba188fc30)

config.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/monster/data/model/TinyLlama-1.1B-Chat-v1.0",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 2,
10
+ "head_dim": 64,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 2048,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 5632,
15
+ "max_position_embeddings": 2048,
16
+ "mlp_bias": false,
17
+ "model_type": "llama",
18
+ "num_attention_heads": 32,
19
+ "num_hidden_layers": 22,
20
+ "num_key_value_heads": 4,
21
+ "pretraining_tp": 1,
22
+ "quantization_config": {
23
+ "bits": 4,
24
+ "checkpoint_format": "gptq",
25
+ "damp_auto_increment": 0.0015,
26
+ "damp_percent": 0.005,
27
+ "desc_act": false,
28
+ "dynamic": null,
29
+ "group_size": 128,
30
+ "lm_head": false,
31
+ "meta": {
32
+ "quantizer": "gptqmodel:1.0.10-dev"
33
+ },
34
+ "model_file_base_name": null,
35
+ "model_name_or_path": null,
36
+ "quant_method": "gptq",
37
+ "static_groups": false,
38
+ "sym": true,
39
+ "true_sequential": true
40
+ },
41
+ "rms_norm_eps": 1e-05,
42
+ "rope_scaling": null,
43
+ "rope_theta": 10000.0,
44
+ "tie_word_embeddings": false,
45
+ "torch_dtype": "bfloat16",
46
+ "transformers_version": "4.45.2",
47
+ "use_cache": true,
48
+ "vocab_size": 32000
49
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:882f363cd8bc71958e4225e56286c7381e6621c0bcb361f22ca78e2a26618ebf
3
+ size 767343824
quant_log.json ADDED
@@ -0,0 +1 @@
 
 
1
+ [{"layer": 0, "module": "self_attn.k_proj", "loss": "0.11426", "damp": "0.00500", "layer_forward_time": "1.138", "w_clone_time": "0.001", "quant_time": "4.205"}, {"layer": 0, "module": "self_attn.v_proj", "loss": "0.00036", "damp": "0.00500", "layer_forward_time": "1.138", "w_clone_time": "0.001", "quant_time": "0.480"}, {"layer": 0, "module": "self_attn.q_proj", "loss": "0.12988", "damp": "0.00500", "layer_forward_time": "1.138", "w_clone_time": "0.006", "quant_time": "0.923"}, {"layer": 0, "module": "self_attn.o_proj", "loss": "0.00002", "damp": "0.00500", "layer_forward_time": "1.320", "w_clone_time": "0.008", "quant_time": "0.306"}, {"layer": 0, "module": "mlp.up_proj", "loss": "0.04541", "damp": "0.00500", "layer_forward_time": "1.335", "w_clone_time": "0.019", "quant_time": "0.407"}, {"layer": 0, "module": "mlp.gate_proj", "loss": "0.05127", "damp": "0.00500", "layer_forward_time": "1.335", "w_clone_time": "0.017", "quant_time": "0.297"}, {"layer": 0, "module": "mlp.down_proj", "loss": "0.00014", "damp": "0.00500", "layer_forward_time": "2.098", "w_clone_time": "0.017", "quant_time": "0.895"}, {"layer": 1, "module": "self_attn.k_proj", "loss": "0.24707", "damp": "0.00500", "layer_forward_time": "1.357", "w_clone_time": "0.006", "quant_time": "0.320"}, {"layer": 1, "module": "self_attn.v_proj", "loss": "0.00188", "damp": "0.00500", "layer_forward_time": "1.357", "w_clone_time": "0.005", "quant_time": "0.555"}, {"layer": 1, "module": "self_attn.q_proj", "loss": "0.34375", "damp": "0.00500", "layer_forward_time": "1.357", "w_clone_time": "0.008", "quant_time": "0.293"}, {"layer": 1, "module": "self_attn.o_proj", "loss": "0.00022", "damp": "0.00500", "layer_forward_time": "0.806", "w_clone_time": "0.006", "quant_time": "0.292"}, {"layer": 1, "module": "mlp.up_proj", "loss": "0.11523", "damp": "0.00500", "layer_forward_time": "0.982", "w_clone_time": "0.020", "quant_time": "0.293"}, {"layer": 1, "module": "mlp.gate_proj", "loss": "0.12891", "damp": "0.00500", "layer_forward_time": "0.982", "w_clone_time": "0.016", "quant_time": "0.295"}, {"layer": 1, "module": "mlp.down_proj", "loss": "0.00048", "damp": "0.00500", "layer_forward_time": "1.846", "w_clone_time": "0.020", "quant_time": "1.021"}, {"layer": 2, "module": "self_attn.k_proj", "loss": "0.16895", "damp": "0.00500", "layer_forward_time": "0.979", "w_clone_time": "0.008", "quant_time": "0.290"}, {"layer": 2, "module": "self_attn.v_proj", "loss": "0.00383", "damp": "0.00500", "layer_forward_time": "0.979", "w_clone_time": "0.001", "quant_time": "0.289"}, {"layer": 2, "module": "self_attn.q_proj", "loss": "0.31445", "damp": "0.00500", "layer_forward_time": "0.979", "w_clone_time": "0.006", "quant_time": "0.292"}, {"layer": 2, "module": "self_attn.o_proj", "loss": "0.00035", "damp": "0.00500", "layer_forward_time": "0.802", "w_clone_time": "0.009", "quant_time": "0.305"}, {"layer": 2, "module": "mlp.up_proj", "loss": "0.21582", "damp": "0.00500", "layer_forward_time": "0.935", "w_clone_time": "0.017", "quant_time": "0.294"}, {"layer": 2, "module": "mlp.gate_proj", "loss": "0.24512", "damp": "0.00500", "layer_forward_time": "0.935", "w_clone_time": "0.017", "quant_time": "0.291"}, {"layer": 2, "module": "mlp.down_proj", "loss": "0.00861", "damp": "0.00500", "layer_forward_time": "1.984", "w_clone_time": "0.019", "quant_time": "0.845"}, {"layer": 3, "module": "self_attn.k_proj", "loss": "0.28516", "damp": "0.00500", "layer_forward_time": "1.040", "w_clone_time": "0.002", "quant_time": "0.312"}, {"layer": 3, "module": "self_attn.v_proj", "loss": "0.01343", "damp": "0.00500", "layer_forward_time": "1.040", "w_clone_time": "0.012", "quant_time": "0.315"}, {"layer": 3, "module": "self_attn.q_proj", "loss": "0.68359", "damp": "0.00500", "layer_forward_time": "1.040", "w_clone_time": "0.010", "quant_time": "0.296"}, {"layer": 3, "module": "self_attn.o_proj", "loss": "0.00032", "damp": "0.00500", "layer_forward_time": "0.813", "w_clone_time": "0.010", "quant_time": "0.297"}, {"layer": 3, "module": "mlp.up_proj", "loss": "0.29688", "damp": "0.00500", "layer_forward_time": "0.912", "w_clone_time": "0.017", "quant_time": "0.295"}, {"layer": 3, "module": "mlp.gate_proj", "loss": "0.34180", "damp": "0.00500", "layer_forward_time": "0.912", "w_clone_time": "0.019", "quant_time": "0.294"}, {"layer": 3, "module": "mlp.down_proj", "loss": "0.00179", "damp": "0.00500", "layer_forward_time": "1.683", "w_clone_time": "0.021", "quant_time": "0.811"}, {"layer": 4, "module": "self_attn.k_proj", "loss": "0.66406", "damp": "0.00500", "layer_forward_time": "1.199", "w_clone_time": "0.001", "quant_time": "0.300"}, {"layer": 4, "module": "self_attn.v_proj", "loss": "0.02271", "damp": "0.00500", "layer_forward_time": "1.199", "w_clone_time": "0.001", "quant_time": "0.290"}, {"layer": 4, "module": "self_attn.q_proj", "loss": "1.33594", "damp": "0.00500", "layer_forward_time": "1.199", "w_clone_time": "0.006", "quant_time": "0.292"}, {"layer": 4, "module": "self_attn.o_proj", "loss": "0.00072", "damp": "0.00500", "layer_forward_time": "0.978", "w_clone_time": "0.006", "quant_time": "0.289"}, {"layer": 4, "module": "mlp.up_proj", "loss": "0.39062", "damp": "0.00500", "layer_forward_time": "1.041", "w_clone_time": "0.017", "quant_time": "0.294"}, {"layer": 4, "module": "mlp.gate_proj", "loss": "0.46289", "damp": "0.00500", "layer_forward_time": "1.041", "w_clone_time": "0.016", "quant_time": "0.328"}, {"layer": 4, "module": "mlp.down_proj", "loss": "0.00275", "damp": "0.00500", "layer_forward_time": "1.793", "w_clone_time": "0.017", "quant_time": "0.830"}, {"layer": 5, "module": "self_attn.k_proj", "loss": "0.45508", "damp": "0.00500", "layer_forward_time": "1.343", "w_clone_time": "0.003", "quant_time": "0.294"}, {"layer": 5, "module": "self_attn.v_proj", "loss": "0.02100", "damp": "0.00500", "layer_forward_time": "1.343", "w_clone_time": "0.009", "quant_time": "0.292"}, {"layer": 5, "module": "self_attn.q_proj", "loss": "0.98828", "damp": "0.00500", "layer_forward_time": "1.343", "w_clone_time": "0.012", "quant_time": "0.297"}, {"layer": 5, "module": "self_attn.o_proj", "loss": "0.00083", "damp": "0.00500", "layer_forward_time": "1.170", "w_clone_time": "0.014", "quant_time": "0.312"}, {"layer": 5, "module": "mlp.up_proj", "loss": "0.48828", "damp": "0.00500", "layer_forward_time": "1.203", "w_clone_time": "0.017", "quant_time": "0.294"}, {"layer": 5, "module": "mlp.gate_proj", "loss": "0.59375", "damp": "0.00500", "layer_forward_time": "1.203", "w_clone_time": "0.020", "quant_time": "0.289"}, {"layer": 5, "module": "mlp.down_proj", "loss": "0.00415", "damp": "0.00500", "layer_forward_time": "1.966", "w_clone_time": "0.017", "quant_time": "0.800"}, {"layer": 6, "module": "self_attn.k_proj", "loss": "0.50391", "damp": "0.00500", "layer_forward_time": "0.968", "w_clone_time": "0.002", "quant_time": "0.290"}, {"layer": 6, "module": "self_attn.v_proj", "loss": "0.02466", "damp": "0.00500", "layer_forward_time": "0.968", "w_clone_time": "0.002", "quant_time": "0.288"}, {"layer": 6, "module": "self_attn.q_proj", "loss": "1.17188", "damp": "0.00500", "layer_forward_time": "0.968", "w_clone_time": "0.006", "quant_time": "0.291"}, {"layer": 6, "module": "self_attn.o_proj", "loss": "0.00123", "damp": "0.00500", "layer_forward_time": "0.792", "w_clone_time": "0.006", "quant_time": "0.293"}, {"layer": 6, "module": "mlp.up_proj", "loss": "0.55469", "damp": "0.00500", "layer_forward_time": "0.936", "w_clone_time": "0.018", "quant_time": "0.295"}, {"layer": 6, "module": "mlp.gate_proj", "loss": "0.71484", "damp": "0.00500", "layer_forward_time": "0.936", "w_clone_time": "0.017", "quant_time": "0.292"}, {"layer": 6, "module": "mlp.down_proj", "loss": "0.00589", "damp": "0.00500", "layer_forward_time": "1.882", "w_clone_time": "0.017", "quant_time": "0.811"}, {"layer": 7, "module": "self_attn.k_proj", "loss": "0.50391", "damp": "0.00500", "layer_forward_time": "1.293", "w_clone_time": "0.006", "quant_time": "0.319"}, {"layer": 7, "module": "self_attn.v_proj", "loss": "0.03735", "damp": "0.00500", "layer_forward_time": "1.293", "w_clone_time": "0.004", "quant_time": "0.291"}, {"layer": 7, "module": "self_attn.q_proj", "loss": "1.42969", "damp": "0.00500", "layer_forward_time": "1.293", "w_clone_time": "0.011", "quant_time": "0.297"}, {"layer": 7, "module": "self_attn.o_proj", "loss": "0.00241", "damp": "0.00500", "layer_forward_time": "1.236", "w_clone_time": "0.011", "quant_time": "0.316"}, {"layer": 7, "module": "mlp.up_proj", "loss": "0.64453", "damp": "0.00500", "layer_forward_time": "1.252", "w_clone_time": "0.057", "quant_time": "0.345"}, {"layer": 7, "module": "mlp.gate_proj", "loss": "0.98438", "damp": "0.00500", "layer_forward_time": "1.252", "w_clone_time": "0.035", "quant_time": "0.339"}, {"layer": 7, "module": "mlp.down_proj", "loss": "0.04834", "damp": "0.00500", "layer_forward_time": "1.976", "w_clone_time": "0.020", "quant_time": "1.075"}, {"layer": 8, "module": "self_attn.k_proj", "loss": "1.05469", "damp": "0.00500", "layer_forward_time": "1.001", "w_clone_time": "0.003", "quant_time": "0.296"}, {"layer": 8, "module": "self_attn.v_proj", "loss": "0.05347", "damp": "0.00500", "layer_forward_time": "1.001", "w_clone_time": "0.001", "quant_time": "0.293"}, {"layer": 8, "module": "self_attn.q_proj", "loss": "2.45312", "damp": "0.00500", "layer_forward_time": "1.001", "w_clone_time": "0.008", "quant_time": "0.300"}, {"layer": 8, "module": "self_attn.o_proj", "loss": "0.00193", "damp": "0.00500", "layer_forward_time": "0.822", "w_clone_time": "0.009", "quant_time": "0.294"}, {"layer": 8, "module": "mlp.up_proj", "loss": "0.78125", "damp": "0.00500", "layer_forward_time": "1.106", "w_clone_time": "0.019", "quant_time": "0.341"}, {"layer": 8, "module": "mlp.gate_proj", "loss": "1.07031", "damp": "0.00500", "layer_forward_time": "1.106", "w_clone_time": "0.016", "quant_time": "0.295"}, {"layer": 8, "module": "mlp.down_proj", "loss": "0.01111", "damp": "0.00500", "layer_forward_time": "1.730", "w_clone_time": "0.017", "quant_time": "0.827"}, {"layer": 9, "module": "self_attn.k_proj", "loss": "0.56250", "damp": "0.00500", "layer_forward_time": "1.176", "w_clone_time": "0.003", "quant_time": "0.292"}, {"layer": 9, "module": "self_attn.v_proj", "loss": "0.03394", "damp": "0.00500", "layer_forward_time": "1.176", "w_clone_time": "0.003", "quant_time": "0.292"}, {"layer": 9, "module": "self_attn.q_proj", "loss": "1.37500", "damp": "0.00500", "layer_forward_time": "1.176", "w_clone_time": "0.010", "quant_time": "0.298"}, {"layer": 9, "module": "self_attn.o_proj", "loss": "0.00430", "damp": "0.00500", "layer_forward_time": "1.077", "w_clone_time": "0.008", "quant_time": "0.296"}, {"layer": 9, "module": "mlp.up_proj", "loss": "0.85156", "damp": "0.00500", "layer_forward_time": "1.106", "w_clone_time": "0.017", "quant_time": "0.298"}, {"layer": 9, "module": "mlp.gate_proj", "loss": "1.24219", "damp": "0.00500", "layer_forward_time": "1.106", "w_clone_time": "0.017", "quant_time": "0.296"}, {"layer": 9, "module": "mlp.down_proj", "loss": "0.01477", "damp": "0.00500", "layer_forward_time": "1.740", "w_clone_time": "0.043", "quant_time": "0.863"}, {"layer": 10, "module": "self_attn.k_proj", "loss": "0.64453", "damp": "0.00500", "layer_forward_time": "0.845", "w_clone_time": "0.005", "quant_time": "0.292"}, {"layer": 10, "module": "self_attn.v_proj", "loss": "0.03760", "damp": "0.00500", "layer_forward_time": "0.845", "w_clone_time": "0.007", "quant_time": "0.297"}, {"layer": 10, "module": "self_attn.q_proj", "loss": "1.51562", "damp": "0.00500", "layer_forward_time": "0.845", "w_clone_time": "0.010", "quant_time": "0.290"}, {"layer": 10, "module": "self_attn.o_proj", "loss": "0.00494", "damp": "0.00500", "layer_forward_time": "0.597", "w_clone_time": "0.015", "quant_time": "0.365"}, {"layer": 10, "module": "mlp.up_proj", "loss": "0.93359", "damp": "0.00500", "layer_forward_time": "0.914", "w_clone_time": "0.020", "quant_time": "0.299"}, {"layer": 10, "module": "mlp.gate_proj", "loss": "1.28906", "damp": "0.00500", "layer_forward_time": "0.914", "w_clone_time": "0.016", "quant_time": "0.296"}, {"layer": 10, "module": "mlp.down_proj", "loss": "0.01672", "damp": "0.00500", "layer_forward_time": "1.816", "w_clone_time": "0.017", "quant_time": "0.830"}, {"layer": 11, "module": "self_attn.k_proj", "loss": "0.77344", "damp": "0.00500", "layer_forward_time": "1.321", "w_clone_time": "0.001", "quant_time": "0.352"}, {"layer": 11, "module": "self_attn.v_proj", "loss": "0.04980", "damp": "0.00500", "layer_forward_time": "1.321", "w_clone_time": "0.001", "quant_time": "0.293"}, {"layer": 11, "module": "self_attn.q_proj", "loss": "2.12500", "damp": "0.00500", "layer_forward_time": "1.321", "w_clone_time": "0.006", "quant_time": "0.294"}, {"layer": 11, "module": "self_attn.o_proj", "loss": "0.00742", "damp": "0.00500", "layer_forward_time": "1.191", "w_clone_time": "0.007", "quant_time": "0.297"}, {"layer": 11, "module": "mlp.up_proj", "loss": "1.05469", "damp": "0.00500", "layer_forward_time": "1.246", "w_clone_time": "0.017", "quant_time": "0.292"}, {"layer": 11, "module": "mlp.gate_proj", "loss": "1.48438", "damp": "0.00500", "layer_forward_time": "1.246", "w_clone_time": "0.016", "quant_time": "0.309"}, {"layer": 11, "module": "mlp.down_proj", "loss": "0.02063", "damp": "0.00500", "layer_forward_time": "1.931", "w_clone_time": "0.027", "quant_time": "0.867"}, {"layer": 12, "module": "self_attn.k_proj", "loss": "0.67188", "damp": "0.00500", "layer_forward_time": "1.168", "w_clone_time": "0.004", "quant_time": "0.293"}, {"layer": 12, "module": "self_attn.v_proj", "loss": "0.05835", "damp": "0.00500", "layer_forward_time": "1.168", "w_clone_time": "0.005", "quant_time": "0.286"}, {"layer": 12, "module": "self_attn.q_proj", "loss": "1.60156", "damp": "0.00500", "layer_forward_time": "1.168", "w_clone_time": "0.008", "quant_time": "0.288"}, {"layer": 12, "module": "self_attn.o_proj", "loss": "0.00964", "damp": "0.00500", "layer_forward_time": "0.972", "w_clone_time": "0.006", "quant_time": "0.291"}, {"layer": 12, "module": "mlp.up_proj", "loss": "1.17969", "damp": "0.00500", "layer_forward_time": "1.045", "w_clone_time": "0.018", "quant_time": "0.307"}, {"layer": 12, "module": "mlp.gate_proj", "loss": "1.81250", "damp": "0.00500", "layer_forward_time": "1.045", "w_clone_time": "0.018", "quant_time": "0.291"}, {"layer": 12, "module": "mlp.down_proj", "loss": "0.02832", "damp": "0.00500", "layer_forward_time": "1.779", "w_clone_time": "0.020", "quant_time": "0.805"}, {"layer": 13, "module": "self_attn.k_proj", "loss": "0.77344", "damp": "0.00500", "layer_forward_time": "1.129", "w_clone_time": "0.001", "quant_time": "0.286"}, {"layer": 13, "module": "self_attn.v_proj", "loss": "0.04980", "damp": "0.00500", "layer_forward_time": "1.129", "w_clone_time": "0.005", "quant_time": "0.297"}, {"layer": 13, "module": "self_attn.q_proj", "loss": "1.78906", "damp": "0.00500", "layer_forward_time": "1.129", "w_clone_time": "0.006", "quant_time": "0.292"}, {"layer": 13, "module": "self_attn.o_proj", "loss": "0.01227", "damp": "0.00500", "layer_forward_time": "0.919", "w_clone_time": "0.006", "quant_time": "0.290"}, {"layer": 13, "module": "mlp.up_proj", "loss": "1.32031", "damp": "0.00500", "layer_forward_time": "1.111", "w_clone_time": "0.019", "quant_time": "0.342"}, {"layer": 13, "module": "mlp.gate_proj", "loss": "1.97656", "damp": "0.00500", "layer_forward_time": "1.111", "w_clone_time": "0.017", "quant_time": "0.300"}, {"layer": 13, "module": "mlp.down_proj", "loss": "0.03735", "damp": "0.00500", "layer_forward_time": "1.759", "w_clone_time": "0.017", "quant_time": "0.818"}, {"layer": 14, "module": "self_attn.k_proj", "loss": "0.75000", "damp": "0.00500", "layer_forward_time": "1.041", "w_clone_time": "0.001", "quant_time": "0.290"}, {"layer": 14, "module": "self_attn.v_proj", "loss": "0.05640", "damp": "0.00500", "layer_forward_time": "1.041", "w_clone_time": "0.001", "quant_time": "0.294"}, {"layer": 14, "module": "self_attn.q_proj", "loss": "1.68750", "damp": "0.00500", "layer_forward_time": "1.041", "w_clone_time": "0.006", "quant_time": "0.291"}, {"layer": 14, "module": "self_attn.o_proj", "loss": "0.01917", "damp": "0.00500", "layer_forward_time": "0.863", "w_clone_time": "0.006", "quant_time": "0.295"}, {"layer": 14, "module": "mlp.up_proj", "loss": "1.48438", "damp": "0.00500", "layer_forward_time": "0.893", "w_clone_time": "0.017", "quant_time": "0.300"}, {"layer": 14, "module": "mlp.gate_proj", "loss": "2.12500", "damp": "0.00500", "layer_forward_time": "0.893", "w_clone_time": "0.018", "quant_time": "0.294"}, {"layer": 14, "module": "mlp.down_proj", "loss": "0.04810", "damp": "0.00500", "layer_forward_time": "1.451", "w_clone_time": "0.017", "quant_time": "0.819"}, {"layer": 15, "module": "self_attn.k_proj", "loss": "0.82422", "damp": "0.00500", "layer_forward_time": "1.060", "w_clone_time": "0.003", "quant_time": "0.293"}, {"layer": 15, "module": "self_attn.v_proj", "loss": "0.09668", "damp": "0.00500", "layer_forward_time": "1.060", "w_clone_time": "0.005", "quant_time": "0.300"}, {"layer": 15, "module": "self_attn.q_proj", "loss": "2.46875", "damp": "0.00500", "layer_forward_time": "1.060", "w_clone_time": "0.009", "quant_time": "0.302"}, {"layer": 15, "module": "self_attn.o_proj", "loss": "0.01709", "damp": "0.00500", "layer_forward_time": "0.923", "w_clone_time": "0.009", "quant_time": "0.289"}, {"layer": 15, "module": "mlp.up_proj", "loss": "1.70312", "damp": "0.00500", "layer_forward_time": "1.146", "w_clone_time": "0.017", "quant_time": "0.292"}, {"layer": 15, "module": "mlp.gate_proj", "loss": "2.35938", "damp": "0.00500", "layer_forward_time": "1.146", "w_clone_time": "0.016", "quant_time": "0.292"}, {"layer": 15, "module": "mlp.down_proj", "loss": "0.07666", "damp": "0.00500", "layer_forward_time": "2.029", "w_clone_time": "0.017", "quant_time": "0.814"}, {"layer": 16, "module": "self_attn.k_proj", "loss": "0.83594", "damp": "0.00500", "layer_forward_time": "0.995", "w_clone_time": "0.005", "quant_time": "0.296"}, {"layer": 16, "module": "self_attn.v_proj", "loss": "0.08936", "damp": "0.00500", "layer_forward_time": "0.995", "w_clone_time": "0.006", "quant_time": "0.287"}, {"layer": 16, "module": "self_attn.q_proj", "loss": "2.48438", "damp": "0.00500", "layer_forward_time": "0.995", "w_clone_time": "0.006", "quant_time": "0.288"}, {"layer": 16, "module": "self_attn.o_proj", "loss": "0.02112", "damp": "0.00500", "layer_forward_time": "0.811", "w_clone_time": "0.006", "quant_time": "0.288"}, {"layer": 16, "module": "mlp.up_proj", "loss": "2.04688", "damp": "0.00500", "layer_forward_time": "0.925", "w_clone_time": "0.017", "quant_time": "0.289"}, {"layer": 16, "module": "mlp.gate_proj", "loss": "3.00000", "damp": "0.00500", "layer_forward_time": "0.925", "w_clone_time": "0.017", "quant_time": "0.289"}, {"layer": 16, "module": "mlp.down_proj", "loss": "0.12354", "damp": "0.00500", "layer_forward_time": "1.947", "w_clone_time": "0.017", "quant_time": "0.818"}, {"layer": 17, "module": "self_attn.k_proj", "loss": "0.86328", "damp": "0.00500", "layer_forward_time": "1.347", "w_clone_time": "0.007", "quant_time": "0.288"}, {"layer": 17, "module": "self_attn.v_proj", "loss": "0.16895", "damp": "0.00500", "layer_forward_time": "1.347", "w_clone_time": "0.003", "quant_time": "0.289"}, {"layer": 17, "module": "self_attn.q_proj", "loss": "2.29688", "damp": "0.00500", "layer_forward_time": "1.347", "w_clone_time": "0.008", "quant_time": "0.289"}, {"layer": 17, "module": "self_attn.o_proj", "loss": "0.02808", "damp": "0.00500", "layer_forward_time": "1.066", "w_clone_time": "0.007", "quant_time": "0.288"}, {"layer": 17, "module": "mlp.up_proj", "loss": "2.43750", "damp": "0.00500", "layer_forward_time": "1.152", "w_clone_time": "0.017", "quant_time": "0.293"}, {"layer": 17, "module": "mlp.gate_proj", "loss": "3.45312", "damp": "0.00500", "layer_forward_time": "1.152", "w_clone_time": "0.018", "quant_time": "0.293"}, {"layer": 17, "module": "mlp.down_proj", "loss": "0.14453", "damp": "0.00500", "layer_forward_time": "1.864", "w_clone_time": "0.018", "quant_time": "0.821"}, {"layer": 18, "module": "self_attn.k_proj", "loss": "0.98828", "damp": "0.00500", "layer_forward_time": "1.161", "w_clone_time": "0.007", "quant_time": "0.285"}, {"layer": 18, "module": "self_attn.v_proj", "loss": "0.26758", "damp": "0.00500", "layer_forward_time": "1.161", "w_clone_time": "0.005", "quant_time": "0.290"}, {"layer": 18, "module": "self_attn.q_proj", "loss": "2.57812", "damp": "0.00500", "layer_forward_time": "1.161", "w_clone_time": "0.006", "quant_time": "0.296"}, {"layer": 18, "module": "self_attn.o_proj", "loss": "0.03882", "damp": "0.00500", "layer_forward_time": "0.992", "w_clone_time": "0.006", "quant_time": "0.287"}, {"layer": 18, "module": "mlp.up_proj", "loss": "3.00000", "damp": "0.00500", "layer_forward_time": "1.019", "w_clone_time": "0.023", "quant_time": "0.298"}, {"layer": 18, "module": "mlp.gate_proj", "loss": "4.15625", "damp": "0.00500", "layer_forward_time": "1.019", "w_clone_time": "0.019", "quant_time": "0.291"}, {"layer": 18, "module": "mlp.down_proj", "loss": "0.22949", "damp": "0.00500", "layer_forward_time": "1.973", "w_clone_time": "0.020", "quant_time": "0.825"}, {"layer": 19, "module": "self_attn.k_proj", "loss": "0.90234", "damp": "0.00500", "layer_forward_time": "1.290", "w_clone_time": "0.002", "quant_time": "0.327"}, {"layer": 19, "module": "self_attn.v_proj", "loss": "0.29883", "damp": "0.00500", "layer_forward_time": "1.290", "w_clone_time": "0.002", "quant_time": "0.290"}, {"layer": 19, "module": "self_attn.q_proj", "loss": "2.42188", "damp": "0.00500", "layer_forward_time": "1.290", "w_clone_time": "0.006", "quant_time": "0.291"}, {"layer": 19, "module": "self_attn.o_proj", "loss": "0.05127", "damp": "0.00500", "layer_forward_time": "1.205", "w_clone_time": "0.010", "quant_time": "0.286"}, {"layer": 19, "module": "mlp.up_proj", "loss": "3.67188", "damp": "0.00500", "layer_forward_time": "1.131", "w_clone_time": "0.020", "quant_time": "0.286"}, {"layer": 19, "module": "mlp.gate_proj", "loss": "4.90625", "damp": "0.00500", "layer_forward_time": "1.131", "w_clone_time": "0.017", "quant_time": "0.297"}, {"layer": 19, "module": "mlp.down_proj", "loss": "0.37305", "damp": "0.00500", "layer_forward_time": "1.896", "w_clone_time": "0.018", "quant_time": "0.811"}, {"layer": 20, "module": "self_attn.k_proj", "loss": "0.90625", "damp": "0.00500", "layer_forward_time": "1.261", "w_clone_time": "0.002", "quant_time": "0.346"}, {"layer": 20, "module": "self_attn.v_proj", "loss": "0.23438", "damp": "0.00500", "layer_forward_time": "1.261", "w_clone_time": "0.001", "quant_time": "0.292"}, {"layer": 20, "module": "self_attn.q_proj", "loss": "2.48438", "damp": "0.00500", "layer_forward_time": "1.261", "w_clone_time": "0.009", "quant_time": "0.290"}, {"layer": 20, "module": "self_attn.o_proj", "loss": "0.06836", "damp": "0.00500", "layer_forward_time": "0.969", "w_clone_time": "0.009", "quant_time": "0.295"}, {"layer": 20, "module": "mlp.up_proj", "loss": "4.34375", "damp": "0.00500", "layer_forward_time": "1.141", "w_clone_time": "0.020", "quant_time": "0.315"}, {"layer": 20, "module": "mlp.gate_proj", "loss": "5.62500", "damp": "0.00500", "layer_forward_time": "1.141", "w_clone_time": "0.019", "quant_time": "0.286"}, {"layer": 20, "module": "mlp.down_proj", "loss": "0.59375", "damp": "0.00500", "layer_forward_time": "1.670", "w_clone_time": "0.017", "quant_time": "0.804"}, {"layer": 21, "module": "self_attn.k_proj", "loss": "0.93359", "damp": "0.00500", "layer_forward_time": "0.901", "w_clone_time": "0.002", "quant_time": "0.285"}, {"layer": 21, "module": "self_attn.v_proj", "loss": "0.31250", "damp": "0.00500", "layer_forward_time": "0.901", "w_clone_time": "0.001", "quant_time": "0.293"}, {"layer": 21, "module": "self_attn.q_proj", "loss": "2.68750", "damp": "0.00500", "layer_forward_time": "0.901", "w_clone_time": "0.008", "quant_time": "0.293"}, {"layer": 21, "module": "self_attn.o_proj", "loss": "0.11377", "damp": "0.00500", "layer_forward_time": "0.770", "w_clone_time": "0.014", "quant_time": "0.303"}, {"layer": 21, "module": "mlp.up_proj", "loss": "4.65625", "damp": "0.00500", "layer_forward_time": "0.870", "w_clone_time": "0.023", "quant_time": "0.289"}, {"layer": 21, "module": "mlp.gate_proj", "loss": "7.34375", "damp": "0.00500", "layer_forward_time": "0.870", "w_clone_time": "0.019", "quant_time": "0.287"}, {"layer": 21, "module": "mlp.down_proj", "loss": "1.06250", "damp": "0.00500", "layer_forward_time": "1.509", "w_clone_time": "0.018", "quant_time": "0.808"}]
quantize_config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 4,
3
+ "dynamic": null,
4
+ "group_size": 128,
5
+ "desc_act": false,
6
+ "static_groups": false,
7
+ "sym": true,
8
+ "lm_head": false,
9
+ "damp_percent": 0.005,
10
+ "damp_auto_increment": 0.0015,
11
+ "true_sequential": true,
12
+ "model_name_or_path": "/monster/data/model/TinyLlama-1.1B-Chat-v1.0/gptq_4bits_10-24_10-47-05_maxlen2048_ns1024_descFalse_damp0.005",
13
+ "model_file_base_name": "model",
14
+ "quant_method": "gptq",
15
+ "checkpoint_format": "gptq",
16
+ "meta": {
17
+ "quantizer": "gptqmodel:1.0.10-dev"
18
+ }
19
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "bos_token": "<s>",
32
+ "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
33
+ "clean_up_tokenization_spaces": false,
34
+ "eos_token": "</s>",
35
+ "legacy": false,
36
+ "model_max_length": 2048,
37
+ "pad_token": "</s>",
38
+ "padding_side": "right",
39
+ "sp_model_kwargs": {},
40
+ "tokenizer_class": "LlamaTokenizer",
41
+ "unk_token": "<unk>",
42
+ "use_default_system_prompt": false
43
+ }