ZeroUniqueness commited on
Commit
b99e125
1 Parent(s): 63a4e86

Training in progress, step 4300

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. checkpoint-4200/README.md +20 -0
  2. checkpoint-4200/adapter_config.json +26 -0
  3. checkpoint-4200/adapter_model.bin +3 -0
  4. checkpoint-4200/adapter_model/README.md +20 -0
  5. checkpoint-4200/adapter_model/adapter_config.json +26 -0
  6. checkpoint-4200/adapter_model/adapter_model.bin +3 -0
  7. checkpoint-4200/optimizer.pt +3 -0
  8. checkpoint-4200/rng_state_0.pth +3 -0
  9. checkpoint-4200/rng_state_1.pth +3 -0
  10. checkpoint-4200/rng_state_10.pth +3 -0
  11. checkpoint-4200/rng_state_11.pth +3 -0
  12. checkpoint-4200/rng_state_12.pth +3 -0
  13. checkpoint-4200/rng_state_13.pth +3 -0
  14. checkpoint-4200/rng_state_2.pth +3 -0
  15. checkpoint-4200/rng_state_3.pth +3 -0
  16. checkpoint-4200/rng_state_4.pth +3 -0
  17. checkpoint-4200/rng_state_5.pth +3 -0
  18. checkpoint-4200/rng_state_6.pth +3 -0
  19. checkpoint-4200/rng_state_7.pth +3 -0
  20. checkpoint-4200/rng_state_8.pth +3 -0
  21. checkpoint-4200/rng_state_9.pth +3 -0
  22. checkpoint-4200/scheduler.pt +3 -0
  23. checkpoint-4200/trainer_state.json +1032 -0
  24. checkpoint-4200/training_args.bin +3 -0
  25. checkpoint-4300/README.md +20 -0
  26. checkpoint-4300/adapter_config.json +26 -0
  27. checkpoint-4300/adapter_model.bin +3 -0
  28. checkpoint-4300/adapter_model/README.md +20 -0
  29. checkpoint-4300/adapter_model/adapter_config.json +26 -0
  30. checkpoint-4300/adapter_model/adapter_model.bin +3 -0
  31. checkpoint-4300/optimizer.pt +3 -0
  32. checkpoint-4300/rng_state_0.pth +3 -0
  33. checkpoint-4300/rng_state_1.pth +3 -0
  34. checkpoint-4300/rng_state_10.pth +3 -0
  35. checkpoint-4300/rng_state_11.pth +3 -0
  36. checkpoint-4300/rng_state_12.pth +3 -0
  37. checkpoint-4300/rng_state_13.pth +3 -0
  38. checkpoint-4300/rng_state_2.pth +3 -0
  39. checkpoint-4300/rng_state_3.pth +3 -0
  40. checkpoint-4300/rng_state_4.pth +3 -0
  41. checkpoint-4300/rng_state_5.pth +3 -0
  42. checkpoint-4300/rng_state_6.pth +3 -0
  43. checkpoint-4300/rng_state_7.pth +3 -0
  44. checkpoint-4300/rng_state_8.pth +3 -0
  45. checkpoint-4300/rng_state_9.pth +3 -0
  46. checkpoint-4300/scheduler.pt +3 -0
  47. checkpoint-4300/trainer_state.json +1056 -0
  48. checkpoint-4300/training_args.bin +3 -0
  49. checkpoint-4400/README.md +20 -0
  50. checkpoint-4400/adapter_config.json +26 -0
checkpoint-4200/README.md ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - load_in_8bit: False
9
+ - load_in_4bit: True
10
+ - llm_int8_threshold: 6.0
11
+ - llm_int8_skip_modules: None
12
+ - llm_int8_enable_fp32_cpu_offload: False
13
+ - llm_int8_has_fp16_weight: False
14
+ - bnb_4bit_quant_type: nf4
15
+ - bnb_4bit_use_double_quant: True
16
+ - bnb_4bit_compute_dtype: bfloat16
17
+ ### Framework versions
18
+
19
+
20
+ - PEFT 0.5.0.dev0
checkpoint-4200/adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "/workspace/webui/models/TheBloke_Llama-2-13B-fp16",
4
+ "bias": "none",
5
+ "fan_in_fan_out": null,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 16,
11
+ "lora_dropout": 0.05,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 32,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "down_proj",
18
+ "o_proj",
19
+ "v_proj",
20
+ "k_proj",
21
+ "up_proj",
22
+ "gate_proj",
23
+ "q_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM"
26
+ }
checkpoint-4200/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3e333df20d090244c536ca681f370106e84c45d6f6033e42293d5dd935148eb
3
+ size 500897101
checkpoint-4200/adapter_model/README.md ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - load_in_8bit: False
9
+ - load_in_4bit: True
10
+ - llm_int8_threshold: 6.0
11
+ - llm_int8_skip_modules: None
12
+ - llm_int8_enable_fp32_cpu_offload: False
13
+ - llm_int8_has_fp16_weight: False
14
+ - bnb_4bit_quant_type: nf4
15
+ - bnb_4bit_use_double_quant: True
16
+ - bnb_4bit_compute_dtype: bfloat16
17
+ ### Framework versions
18
+
19
+
20
+ - PEFT 0.5.0.dev0
checkpoint-4200/adapter_model/adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "/workspace/webui/models/TheBloke_Llama-2-13B-fp16",
4
+ "bias": "none",
5
+ "fan_in_fan_out": null,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 16,
11
+ "lora_dropout": 0.05,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 32,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "down_proj",
18
+ "o_proj",
19
+ "v_proj",
20
+ "k_proj",
21
+ "up_proj",
22
+ "gate_proj",
23
+ "q_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM"
26
+ }
checkpoint-4200/adapter_model/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3e333df20d090244c536ca681f370106e84c45d6f6033e42293d5dd935148eb
3
+ size 500897101
checkpoint-4200/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77f2eb217cfb36ea105abf27909c020d1f994a2f0d0b9529c4f34c95fa1979a6
3
+ size 1001752701
checkpoint-4200/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e39955e63618e46f3bdc48794eb468c4a2a362a14b852408614167c948270a9f
3
+ size 27772
checkpoint-4200/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ace003ac134e276924eb3ea4f065dababf4315826919aa87c57bd6ed1f3c9f8f
3
+ size 27772
checkpoint-4200/rng_state_10.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7214f06b466b42d304a3bac986a1c2ef8dc14dafe3ca963c50cc13faf825db9
3
+ size 27789
checkpoint-4200/rng_state_11.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9321a3c8b3f28ea39e71d0c9199f896d1989599a3945fb573f479f28b13c7081
3
+ size 27789
checkpoint-4200/rng_state_12.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0adc7008fcd6248e4886234c2141a70e79e8b53bf5fc19aae0deb1b8a5ffbdbd
3
+ size 27789
checkpoint-4200/rng_state_13.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e9ca26dbc3ab046b9d5dedf18e3e181aa09c97d2ad441ced93c7a1d82f73fca
3
+ size 27789
checkpoint-4200/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bbe635548edc83c591c14a2581c100e77475547bba010c7efba7f187a428d46
3
+ size 27772
checkpoint-4200/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:47f9eaccfe6fd830f82205d0a5f025c6394d5f803d79213cd73210a58146d9b0
3
+ size 27772
checkpoint-4200/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:340adbcca40d74a681243ee67ff7bc6c95c55289c128f515419588af41856bcf
3
+ size 27772
checkpoint-4200/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc872f2e697b2a05a8fdeaaef5d8a5579254541515d2738e64317b15f0aea5be
3
+ size 27772
checkpoint-4200/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d80328e735a3525dddccf939e91d6eca35158a45e34301c42a8db036b696176b
3
+ size 27772
checkpoint-4200/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f1edc9d338e7a89fd54ffa70b9782ed3c0abe6fbf326b9254957a771c5923e8
3
+ size 27772
checkpoint-4200/rng_state_8.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:514fbb84e63510c3d0922a39c6f9a4922b21b3dde59c727d229088232a708135
3
+ size 27772
checkpoint-4200/rng_state_9.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1ee58ff754f86aeb8212b1e7ab704ad3b725ce3348aad288674dee695e18524
3
+ size 27772
checkpoint-4200/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03d7576e67497d9cef602df64ab49be9ebeda7e0cb4f0d510989a98dd6dae490
3
+ size 627
checkpoint-4200/trainer_state.json ADDED
@@ -0,0 +1,1032 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.6285381930981,
5
+ "global_step": 4200,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.02,
12
+ "learning_rate": 0.0001999867761371633,
13
+ "loss": 1.0435,
14
+ "step": 50
15
+ },
16
+ {
17
+ "epoch": 0.04,
18
+ "learning_rate": 0.00019993306018843102,
19
+ "loss": 0.8918,
20
+ "step": 100
21
+ },
22
+ {
23
+ "epoch": 0.06,
24
+ "learning_rate": 0.00019983804784290833,
25
+ "loss": 0.8874,
26
+ "step": 150
27
+ },
28
+ {
29
+ "epoch": 0.08,
30
+ "learning_rate": 0.00019970177836355307,
31
+ "loss": 0.8839,
32
+ "step": 200
33
+ },
34
+ {
35
+ "epoch": 0.09,
36
+ "learning_rate": 0.00019961818913082012,
37
+ "loss": 0.8801,
38
+ "step": 225
39
+ },
40
+ {
41
+ "epoch": 0.1,
42
+ "learning_rate": 0.00019952430806244534,
43
+ "loss": 0.8753,
44
+ "step": 250
45
+ },
46
+ {
47
+ "epoch": 0.11,
48
+ "learning_rate": 0.00019942014485754635,
49
+ "loss": 0.8754,
50
+ "step": 275
51
+ },
52
+ {
53
+ "epoch": 0.12,
54
+ "learning_rate": 0.00019930571027751713,
55
+ "loss": 0.8751,
56
+ "step": 300
57
+ },
58
+ {
59
+ "epoch": 0.13,
60
+ "learning_rate": 0.0001991810161449164,
61
+ "loss": 0.8819,
62
+ "step": 325
63
+ },
64
+ {
65
+ "epoch": 0.14,
66
+ "learning_rate": 0.00019904607534224612,
67
+ "loss": 0.8744,
68
+ "step": 350
69
+ },
70
+ {
71
+ "epoch": 0.15,
72
+ "learning_rate": 0.00019890090181062063,
73
+ "loss": 0.8735,
74
+ "step": 375
75
+ },
76
+ {
77
+ "epoch": 0.16,
78
+ "learning_rate": 0.00019874551054832625,
79
+ "loss": 0.8703,
80
+ "step": 400
81
+ },
82
+ {
83
+ "epoch": 0.16,
84
+ "learning_rate": 0.00019857991760927193,
85
+ "loss": 0.8715,
86
+ "step": 425
87
+ },
88
+ {
89
+ "epoch": 0.17,
90
+ "learning_rate": 0.00019840414010133045,
91
+ "loss": 0.8714,
92
+ "step": 450
93
+ },
94
+ {
95
+ "epoch": 0.18,
96
+ "learning_rate": 0.00019821819618457114,
97
+ "loss": 0.8653,
98
+ "step": 475
99
+ },
100
+ {
101
+ "epoch": 0.19,
102
+ "learning_rate": 0.0001980221050693837,
103
+ "loss": 0.8716,
104
+ "step": 500
105
+ },
106
+ {
107
+ "epoch": 0.2,
108
+ "learning_rate": 0.00019781588701449338,
109
+ "loss": 0.8695,
110
+ "step": 525
111
+ },
112
+ {
113
+ "epoch": 0.21,
114
+ "learning_rate": 0.0001975995633248682,
115
+ "loss": 0.8746,
116
+ "step": 550
117
+ },
118
+ {
119
+ "epoch": 0.22,
120
+ "learning_rate": 0.00019737315634951762,
121
+ "loss": 0.8731,
122
+ "step": 575
123
+ },
124
+ {
125
+ "epoch": 0.23,
126
+ "learning_rate": 0.00019713668947918386,
127
+ "loss": 0.867,
128
+ "step": 600
129
+ },
130
+ {
131
+ "epoch": 0.24,
132
+ "learning_rate": 0.0001968901871439252,
133
+ "loss": 0.8706,
134
+ "step": 625
135
+ },
136
+ {
137
+ "epoch": 0.25,
138
+ "learning_rate": 0.000196633674810592,
139
+ "loss": 0.8595,
140
+ "step": 650
141
+ },
142
+ {
143
+ "epoch": 0.26,
144
+ "learning_rate": 0.0001963671789801958,
145
+ "loss": 0.8627,
146
+ "step": 675
147
+ },
148
+ {
149
+ "epoch": 0.27,
150
+ "learning_rate": 0.0001960907271851712,
151
+ "loss": 0.8607,
152
+ "step": 700
153
+ },
154
+ {
155
+ "epoch": 0.28,
156
+ "learning_rate": 0.00019580434798653173,
157
+ "loss": 0.858,
158
+ "step": 725
159
+ },
160
+ {
161
+ "epoch": 0.29,
162
+ "learning_rate": 0.00019550807097091876,
163
+ "loss": 0.8589,
164
+ "step": 750
165
+ },
166
+ {
167
+ "epoch": 0.3,
168
+ "learning_rate": 0.00019520192674754515,
169
+ "loss": 0.8561,
170
+ "step": 775
171
+ },
172
+ {
173
+ "epoch": 0.31,
174
+ "learning_rate": 0.00019488594694503264,
175
+ "loss": 0.8576,
176
+ "step": 800
177
+ },
178
+ {
179
+ "epoch": 0.32,
180
+ "learning_rate": 0.00019456016420814446,
181
+ "loss": 0.8597,
182
+ "step": 825
183
+ },
184
+ {
185
+ "epoch": 0.33,
186
+ "learning_rate": 0.00019422461219441254,
187
+ "loss": 0.862,
188
+ "step": 850
189
+ },
190
+ {
191
+ "epoch": 0.34,
192
+ "learning_rate": 0.00019387932557066035,
193
+ "loss": 0.8577,
194
+ "step": 875
195
+ },
196
+ {
197
+ "epoch": 0.35,
198
+ "learning_rate": 0.00019352434000942127,
199
+ "loss": 0.8632,
200
+ "step": 900
201
+ },
202
+ {
203
+ "epoch": 0.36,
204
+ "learning_rate": 0.00019315969218525333,
205
+ "loss": 0.8567,
206
+ "step": 925
207
+ },
208
+ {
209
+ "epoch": 0.37,
210
+ "learning_rate": 0.00019278541977095005,
211
+ "loss": 0.8501,
212
+ "step": 950
213
+ },
214
+ {
215
+ "epoch": 0.38,
216
+ "learning_rate": 0.00019240156143364844,
217
+ "loss": 0.8596,
218
+ "step": 975
219
+ },
220
+ {
221
+ "epoch": 0.39,
222
+ "learning_rate": 0.00019200815683083434,
223
+ "loss": 0.8556,
224
+ "step": 1000
225
+ },
226
+ {
227
+ "epoch": 0.39,
228
+ "eval_loss": 0.8521950244903564,
229
+ "eval_runtime": 59.8838,
230
+ "eval_samples_per_second": 12.19,
231
+ "eval_steps_per_second": 0.885,
232
+ "step": 1000
233
+ },
234
+ {
235
+ "epoch": 0.4,
236
+ "learning_rate": 0.00019160524660624505,
237
+ "loss": 0.8531,
238
+ "step": 1025
239
+ },
240
+ {
241
+ "epoch": 0.41,
242
+ "learning_rate": 0.00019119287238567045,
243
+ "loss": 0.8513,
244
+ "step": 1050
245
+ },
246
+ {
247
+ "epoch": 0.42,
248
+ "learning_rate": 0.00019077107677265253,
249
+ "loss": 0.8502,
250
+ "step": 1075
251
+ },
252
+ {
253
+ "epoch": 0.43,
254
+ "learning_rate": 0.00019033990334408384,
255
+ "loss": 0.8469,
256
+ "step": 1100
257
+ },
258
+ {
259
+ "epoch": 0.44,
260
+ "learning_rate": 0.00018989939664570545,
261
+ "loss": 0.8495,
262
+ "step": 1125
263
+ },
264
+ {
265
+ "epoch": 0.45,
266
+ "learning_rate": 0.00018944960218750484,
267
+ "loss": 0.8485,
268
+ "step": 1150
269
+ },
270
+ {
271
+ "epoch": 0.46,
272
+ "learning_rate": 0.00018899056643901404,
273
+ "loss": 0.8534,
274
+ "step": 1175
275
+ },
276
+ {
277
+ "epoch": 0.47,
278
+ "learning_rate": 0.00018852233682450893,
279
+ "loss": 0.8531,
280
+ "step": 1200
281
+ },
282
+ {
283
+ "epoch": 0.47,
284
+ "learning_rate": 0.00018804496171810948,
285
+ "loss": 0.8509,
286
+ "step": 1225
287
+ },
288
+ {
289
+ "epoch": 0.48,
290
+ "learning_rate": 0.00018755849043878222,
291
+ "loss": 0.8445,
292
+ "step": 1250
293
+ },
294
+ {
295
+ "epoch": 0.49,
296
+ "learning_rate": 0.0001870629732452449,
297
+ "loss": 0.8548,
298
+ "step": 1275
299
+ },
300
+ {
301
+ "epoch": 0.5,
302
+ "learning_rate": 0.00018655846133077417,
303
+ "loss": 0.8441,
304
+ "step": 1300
305
+ },
306
+ {
307
+ "epoch": 0.51,
308
+ "learning_rate": 0.00018604500681791656,
309
+ "loss": 0.8533,
310
+ "step": 1325
311
+ },
312
+ {
313
+ "epoch": 0.52,
314
+ "learning_rate": 0.00018552266275310373,
315
+ "loss": 0.8505,
316
+ "step": 1350
317
+ },
318
+ {
319
+ "epoch": 0.53,
320
+ "learning_rate": 0.0001849914831011719,
321
+ "loss": 0.8544,
322
+ "step": 1375
323
+ },
324
+ {
325
+ "epoch": 0.54,
326
+ "learning_rate": 0.00018445152273978668,
327
+ "loss": 0.845,
328
+ "step": 1400
329
+ },
330
+ {
331
+ "epoch": 0.55,
332
+ "learning_rate": 0.00018390283745377354,
333
+ "loss": 0.8376,
334
+ "step": 1425
335
+ },
336
+ {
337
+ "epoch": 0.56,
338
+ "learning_rate": 0.0001833454839293545,
339
+ "loss": 0.847,
340
+ "step": 1450
341
+ },
342
+ {
343
+ "epoch": 0.57,
344
+ "learning_rate": 0.00018277951974829163,
345
+ "loss": 0.8473,
346
+ "step": 1475
347
+ },
348
+ {
349
+ "epoch": 0.58,
350
+ "learning_rate": 0.0001822050033819382,
351
+ "loss": 0.8438,
352
+ "step": 1500
353
+ },
354
+ {
355
+ "epoch": 0.59,
356
+ "learning_rate": 0.00018162199418519785,
357
+ "loss": 0.8418,
358
+ "step": 1525
359
+ },
360
+ {
361
+ "epoch": 0.6,
362
+ "learning_rate": 0.00018103055239039243,
363
+ "loss": 0.842,
364
+ "step": 1550
365
+ },
366
+ {
367
+ "epoch": 0.61,
368
+ "learning_rate": 0.0001804307391010393,
369
+ "loss": 0.8435,
370
+ "step": 1575
371
+ },
372
+ {
373
+ "epoch": 0.62,
374
+ "learning_rate": 0.00017982261628553842,
375
+ "loss": 0.8349,
376
+ "step": 1600
377
+ },
378
+ {
379
+ "epoch": 0.63,
380
+ "learning_rate": 0.0001792062467707703,
381
+ "loss": 0.8483,
382
+ "step": 1625
383
+ },
384
+ {
385
+ "epoch": 0.64,
386
+ "learning_rate": 0.0001785816942356052,
387
+ "loss": 0.8387,
388
+ "step": 1650
389
+ },
390
+ {
391
+ "epoch": 0.65,
392
+ "learning_rate": 0.00017794902320432429,
393
+ "loss": 0.843,
394
+ "step": 1675
395
+ },
396
+ {
397
+ "epoch": 0.66,
398
+ "learning_rate": 0.00017730829903995333,
399
+ "loss": 0.8424,
400
+ "step": 1700
401
+ },
402
+ {
403
+ "epoch": 0.67,
404
+ "learning_rate": 0.00017665958793751006,
405
+ "loss": 0.8418,
406
+ "step": 1725
407
+ },
408
+ {
409
+ "epoch": 0.68,
410
+ "learning_rate": 0.00017600295691716522,
411
+ "loss": 0.8384,
412
+ "step": 1750
413
+ },
414
+ {
415
+ "epoch": 0.69,
416
+ "learning_rate": 0.00017533847381731856,
417
+ "loss": 0.8445,
418
+ "step": 1775
419
+ },
420
+ {
421
+ "epoch": 0.7,
422
+ "learning_rate": 0.00017466620728759033,
423
+ "loss": 0.8446,
424
+ "step": 1800
425
+ },
426
+ {
427
+ "epoch": 0.71,
428
+ "learning_rate": 0.00017398622678172878,
429
+ "loss": 0.838,
430
+ "step": 1825
431
+ },
432
+ {
433
+ "epoch": 0.72,
434
+ "learning_rate": 0.0001732986025504348,
435
+ "loss": 0.8415,
436
+ "step": 1850
437
+ },
438
+ {
439
+ "epoch": 0.73,
440
+ "learning_rate": 0.000172603405634104,
441
+ "loss": 0.8357,
442
+ "step": 1875
443
+ },
444
+ {
445
+ "epoch": 0.74,
446
+ "learning_rate": 0.00017190070785548755,
447
+ "loss": 0.8311,
448
+ "step": 1900
449
+ },
450
+ {
451
+ "epoch": 0.75,
452
+ "learning_rate": 0.0001711905818122717,
453
+ "loss": 0.8333,
454
+ "step": 1925
455
+ },
456
+ {
457
+ "epoch": 0.76,
458
+ "learning_rate": 0.0001704731008695777,
459
+ "loss": 0.8387,
460
+ "step": 1950
461
+ },
462
+ {
463
+ "epoch": 0.77,
464
+ "learning_rate": 0.0001697483391523821,
465
+ "loss": 0.8442,
466
+ "step": 1975
467
+ },
468
+ {
469
+ "epoch": 0.78,
470
+ "learning_rate": 0.00016901637153785885,
471
+ "loss": 0.8399,
472
+ "step": 2000
473
+ },
474
+ {
475
+ "epoch": 0.78,
476
+ "eval_loss": 0.8339959383010864,
477
+ "eval_runtime": 58.5829,
478
+ "eval_samples_per_second": 12.461,
479
+ "eval_steps_per_second": 0.905,
480
+ "step": 2000
481
+ },
482
+ {
483
+ "epoch": 0.79,
484
+ "learning_rate": 0.0001682772736476434,
485
+ "loss": 0.8334,
486
+ "step": 2025
487
+ },
488
+ {
489
+ "epoch": 0.79,
490
+ "learning_rate": 0.0001675311218400201,
491
+ "loss": 0.835,
492
+ "step": 2050
493
+ },
494
+ {
495
+ "epoch": 0.8,
496
+ "learning_rate": 0.00016677799320203332,
497
+ "loss": 0.8368,
498
+ "step": 2075
499
+ },
500
+ {
501
+ "epoch": 0.81,
502
+ "learning_rate": 0.00016601796554152344,
503
+ "loss": 0.8278,
504
+ "step": 2100
505
+ },
506
+ {
507
+ "epoch": 0.82,
508
+ "learning_rate": 0.00016525111737908827,
509
+ "loss": 0.8334,
510
+ "step": 2125
511
+ },
512
+ {
513
+ "epoch": 0.83,
514
+ "learning_rate": 0.00016447752793997096,
515
+ "loss": 0.8416,
516
+ "step": 2150
517
+ },
518
+ {
519
+ "epoch": 0.84,
520
+ "learning_rate": 0.00016369727714587483,
521
+ "loss": 0.8297,
522
+ "step": 2175
523
+ },
524
+ {
525
+ "epoch": 0.85,
526
+ "learning_rate": 0.0001629104456067066,
527
+ "loss": 0.8327,
528
+ "step": 2200
529
+ },
530
+ {
531
+ "epoch": 0.86,
532
+ "learning_rate": 0.00016211711461224825,
533
+ "loss": 0.8324,
534
+ "step": 2225
535
+ },
536
+ {
537
+ "epoch": 0.87,
538
+ "learning_rate": 0.0001613173661237589,
539
+ "loss": 0.8313,
540
+ "step": 2250
541
+ },
542
+ {
543
+ "epoch": 0.88,
544
+ "learning_rate": 0.0001605112827655069,
545
+ "loss": 0.8292,
546
+ "step": 2275
547
+ },
548
+ {
549
+ "epoch": 0.89,
550
+ "learning_rate": 0.0001596989478162339,
551
+ "loss": 0.8334,
552
+ "step": 2300
553
+ },
554
+ {
555
+ "epoch": 0.9,
556
+ "learning_rate": 0.00015888044520055106,
557
+ "loss": 0.8352,
558
+ "step": 2325
559
+ },
560
+ {
561
+ "epoch": 0.91,
562
+ "learning_rate": 0.00015805585948026852,
563
+ "loss": 0.823,
564
+ "step": 2350
565
+ },
566
+ {
567
+ "epoch": 0.92,
568
+ "learning_rate": 0.000157225275845659,
569
+ "loss": 0.8293,
570
+ "step": 2375
571
+ },
572
+ {
573
+ "epoch": 0.93,
574
+ "learning_rate": 0.00015638878010665672,
575
+ "loss": 0.8289,
576
+ "step": 2400
577
+ },
578
+ {
579
+ "epoch": 0.94,
580
+ "learning_rate": 0.00015554645868399205,
581
+ "loss": 0.832,
582
+ "step": 2425
583
+ },
584
+ {
585
+ "epoch": 0.95,
586
+ "learning_rate": 0.00015469839860026308,
587
+ "loss": 0.8294,
588
+ "step": 2450
589
+ },
590
+ {
591
+ "epoch": 0.96,
592
+ "learning_rate": 0.0001538446874709452,
593
+ "loss": 0.8281,
594
+ "step": 2475
595
+ },
596
+ {
597
+ "epoch": 0.97,
598
+ "learning_rate": 0.00015298541349533925,
599
+ "loss": 0.8314,
600
+ "step": 2500
601
+ },
602
+ {
603
+ "epoch": 0.98,
604
+ "learning_rate": 0.00015212066544745926,
605
+ "loss": 0.831,
606
+ "step": 2525
607
+ },
608
+ {
609
+ "epoch": 0.99,
610
+ "learning_rate": 0.00015125053266686124,
611
+ "loss": 0.8319,
612
+ "step": 2550
613
+ },
614
+ {
615
+ "epoch": 1.0,
616
+ "learning_rate": 0.00015037510504941303,
617
+ "loss": 0.8259,
618
+ "step": 2575
619
+ },
620
+ {
621
+ "epoch": 1.01,
622
+ "learning_rate": 0.00014949447303800695,
623
+ "loss": 0.8133,
624
+ "step": 2600
625
+ },
626
+ {
627
+ "epoch": 1.02,
628
+ "learning_rate": 0.00014860872761321593,
629
+ "loss": 0.8139,
630
+ "step": 2625
631
+ },
632
+ {
633
+ "epoch": 1.03,
634
+ "learning_rate": 0.00014771796028389405,
635
+ "loss": 0.804,
636
+ "step": 2650
637
+ },
638
+ {
639
+ "epoch": 1.04,
640
+ "learning_rate": 0.0001468222630777225,
641
+ "loss": 0.8011,
642
+ "step": 2675
643
+ },
644
+ {
645
+ "epoch": 1.05,
646
+ "learning_rate": 0.00014592172853170193,
647
+ "loss": 0.8037,
648
+ "step": 2700
649
+ },
650
+ {
651
+ "epoch": 1.06,
652
+ "learning_rate": 0.00014501644968259212,
653
+ "loss": 0.8063,
654
+ "step": 2725
655
+ },
656
+ {
657
+ "epoch": 1.07,
658
+ "learning_rate": 0.00014410652005730025,
659
+ "loss": 0.8155,
660
+ "step": 2750
661
+ },
662
+ {
663
+ "epoch": 1.08,
664
+ "learning_rate": 0.00014319203366321826,
665
+ "loss": 0.8066,
666
+ "step": 2775
667
+ },
668
+ {
669
+ "epoch": 1.09,
670
+ "learning_rate": 0.0001422730849785107,
671
+ "loss": 0.8091,
672
+ "step": 2800
673
+ },
674
+ {
675
+ "epoch": 1.1,
676
+ "learning_rate": 0.0001413497689423539,
677
+ "loss": 0.8067,
678
+ "step": 2825
679
+ },
680
+ {
681
+ "epoch": 1.11,
682
+ "learning_rate": 0.00014042218094512755,
683
+ "loss": 0.8046,
684
+ "step": 2850
685
+ },
686
+ {
687
+ "epoch": 1.11,
688
+ "learning_rate": 0.00013949041681855985,
689
+ "loss": 0.8053,
690
+ "step": 2875
691
+ },
692
+ {
693
+ "epoch": 1.12,
694
+ "learning_rate": 0.0001385545728258264,
695
+ "loss": 0.8075,
696
+ "step": 2900
697
+ },
698
+ {
699
+ "epoch": 1.13,
700
+ "learning_rate": 0.0001376147456516055,
701
+ "loss": 0.8015,
702
+ "step": 2925
703
+ },
704
+ {
705
+ "epoch": 1.14,
706
+ "learning_rate": 0.00013667103239208903,
707
+ "loss": 0.8016,
708
+ "step": 2950
709
+ },
710
+ {
711
+ "epoch": 1.15,
712
+ "learning_rate": 0.00013572353054495126,
713
+ "loss": 0.8029,
714
+ "step": 2975
715
+ },
716
+ {
717
+ "epoch": 1.16,
718
+ "learning_rate": 0.0001347723379992762,
719
+ "loss": 0.8017,
720
+ "step": 3000
721
+ },
722
+ {
723
+ "epoch": 1.16,
724
+ "eval_loss": 0.8229297995567322,
725
+ "eval_runtime": 59.3398,
726
+ "eval_samples_per_second": 12.302,
727
+ "eval_steps_per_second": 0.893,
728
+ "step": 3000
729
+ },
730
+ {
731
+ "epoch": 1.17,
732
+ "learning_rate": 0.0001338175530254443,
733
+ "loss": 0.8049,
734
+ "step": 3025
735
+ },
736
+ {
737
+ "epoch": 1.18,
738
+ "learning_rate": 0.00013285927426497985,
739
+ "loss": 0.8027,
740
+ "step": 3050
741
+ },
742
+ {
743
+ "epoch": 1.19,
744
+ "learning_rate": 0.00013189760072036008,
745
+ "loss": 0.8028,
746
+ "step": 3075
747
+ },
748
+ {
749
+ "epoch": 1.2,
750
+ "learning_rate": 0.0001309326317447869,
751
+ "loss": 0.8021,
752
+ "step": 3100
753
+ },
754
+ {
755
+ "epoch": 1.21,
756
+ "learning_rate": 0.00012996446703192257,
757
+ "loss": 0.8033,
758
+ "step": 3125
759
+ },
760
+ {
761
+ "epoch": 1.22,
762
+ "learning_rate": 0.00012899320660558986,
763
+ "loss": 0.8016,
764
+ "step": 3150
765
+ },
766
+ {
767
+ "epoch": 1.23,
768
+ "learning_rate": 0.00012801895080943846,
769
+ "loss": 0.7995,
770
+ "step": 3175
771
+ },
772
+ {
773
+ "epoch": 1.24,
774
+ "learning_rate": 0.0001270418002965782,
775
+ "loss": 0.799,
776
+ "step": 3200
777
+ },
778
+ {
779
+ "epoch": 1.25,
780
+ "learning_rate": 0.0001260618560191802,
781
+ "loss": 0.8002,
782
+ "step": 3225
783
+ },
784
+ {
785
+ "epoch": 1.26,
786
+ "learning_rate": 0.00012507921921804717,
787
+ "loss": 0.8068,
788
+ "step": 3250
789
+ },
790
+ {
791
+ "epoch": 1.27,
792
+ "learning_rate": 0.00012409399141215423,
793
+ "loss": 0.8041,
794
+ "step": 3275
795
+ },
796
+ {
797
+ "epoch": 1.28,
798
+ "learning_rate": 0.0001231062743881603,
799
+ "loss": 0.7999,
800
+ "step": 3300
801
+ },
802
+ {
803
+ "epoch": 1.29,
804
+ "learning_rate": 0.0001221161701898926,
805
+ "loss": 0.7995,
806
+ "step": 3325
807
+ },
808
+ {
809
+ "epoch": 1.3,
810
+ "learning_rate": 0.00012112378110780391,
811
+ "loss": 0.7959,
812
+ "step": 3350
813
+ },
814
+ {
815
+ "epoch": 1.31,
816
+ "learning_rate": 0.00012012920966840486,
817
+ "loss": 0.7999,
818
+ "step": 3375
819
+ },
820
+ {
821
+ "epoch": 1.32,
822
+ "learning_rate": 0.00011913255862367151,
823
+ "loss": 0.8016,
824
+ "step": 3400
825
+ },
826
+ {
827
+ "epoch": 1.33,
828
+ "learning_rate": 0.00011813393094042993,
829
+ "loss": 0.7944,
830
+ "step": 3425
831
+ },
832
+ {
833
+ "epoch": 1.34,
834
+ "learning_rate": 0.0001171334297897181,
835
+ "loss": 0.8026,
836
+ "step": 3450
837
+ },
838
+ {
839
+ "epoch": 1.35,
840
+ "learning_rate": 0.00011613115853612734,
841
+ "loss": 0.8004,
842
+ "step": 3475
843
+ },
844
+ {
845
+ "epoch": 1.36,
846
+ "learning_rate": 0.00011512722072712321,
847
+ "loss": 0.7992,
848
+ "step": 3500
849
+ },
850
+ {
851
+ "epoch": 1.37,
852
+ "learning_rate": 0.00011412172008234785,
853
+ "loss": 0.8004,
854
+ "step": 3525
855
+ },
856
+ {
857
+ "epoch": 1.38,
858
+ "learning_rate": 0.0001131147604829043,
859
+ "loss": 0.8009,
860
+ "step": 3550
861
+ },
862
+ {
863
+ "epoch": 1.39,
864
+ "learning_rate": 0.00011210644596062439,
865
+ "loss": 0.7993,
866
+ "step": 3575
867
+ },
868
+ {
869
+ "epoch": 1.4,
870
+ "learning_rate": 0.00011109688068732081,
871
+ "loss": 0.7965,
872
+ "step": 3600
873
+ },
874
+ {
875
+ "epoch": 1.41,
876
+ "learning_rate": 0.00011008616896402482,
877
+ "loss": 0.7991,
878
+ "step": 3625
879
+ },
880
+ {
881
+ "epoch": 1.42,
882
+ "learning_rate": 0.00010907441521021072,
883
+ "loss": 0.8026,
884
+ "step": 3650
885
+ },
886
+ {
887
+ "epoch": 1.42,
888
+ "learning_rate": 0.00010806172395300789,
889
+ "loss": 0.7941,
890
+ "step": 3675
891
+ },
892
+ {
893
+ "epoch": 1.43,
894
+ "learning_rate": 0.00010704819981640186,
895
+ "loss": 0.7989,
896
+ "step": 3700
897
+ },
898
+ {
899
+ "epoch": 1.44,
900
+ "learning_rate": 0.00010603394751042522,
901
+ "loss": 0.7981,
902
+ "step": 3725
903
+ },
904
+ {
905
+ "epoch": 1.45,
906
+ "learning_rate": 0.00010501907182033979,
907
+ "loss": 0.7985,
908
+ "step": 3750
909
+ },
910
+ {
911
+ "epoch": 1.46,
912
+ "learning_rate": 0.000104003677595811,
913
+ "loss": 0.7921,
914
+ "step": 3775
915
+ },
916
+ {
917
+ "epoch": 1.47,
918
+ "learning_rate": 0.00010298786974007555,
919
+ "loss": 0.8012,
920
+ "step": 3800
921
+ },
922
+ {
923
+ "epoch": 1.48,
924
+ "learning_rate": 0.00010197175319910343,
925
+ "loss": 0.7906,
926
+ "step": 3825
927
+ },
928
+ {
929
+ "epoch": 1.49,
930
+ "learning_rate": 0.00010095543295075593,
931
+ "loss": 0.7928,
932
+ "step": 3850
933
+ },
934
+ {
935
+ "epoch": 1.5,
936
+ "learning_rate": 9.993901399393979e-05,
937
+ "loss": 0.8018,
938
+ "step": 3875
939
+ },
940
+ {
941
+ "epoch": 1.51,
942
+ "learning_rate": 9.892260133775968e-05,
943
+ "loss": 0.7991,
944
+ "step": 3900
945
+ },
946
+ {
947
+ "epoch": 1.52,
948
+ "learning_rate": 9.79062999906693e-05,
949
+ "loss": 0.795,
950
+ "step": 3925
951
+ },
952
+ {
953
+ "epoch": 1.53,
954
+ "learning_rate": 9.68902149496227e-05,
955
+ "loss": 0.7977,
956
+ "step": 3950
957
+ },
958
+ {
959
+ "epoch": 1.54,
960
+ "learning_rate": 9.587445118922674e-05,
961
+ "loss": 0.8013,
962
+ "step": 3975
963
+ },
964
+ {
965
+ "epoch": 1.55,
966
+ "learning_rate": 9.485911365089589e-05,
967
+ "loss": 0.7978,
968
+ "step": 4000
969
+ },
970
+ {
971
+ "epoch": 1.55,
972
+ "eval_loss": 0.8142631649971008,
973
+ "eval_runtime": 59.4108,
974
+ "eval_samples_per_second": 12.287,
975
+ "eval_steps_per_second": 0.892,
976
+ "step": 4000
977
+ },
978
+ {
979
+ "epoch": 1.56,
980
+ "learning_rate": 9.384430723201036e-05,
981
+ "loss": 0.7912,
982
+ "step": 4025
983
+ },
984
+ {
985
+ "epoch": 1.57,
986
+ "learning_rate": 9.283013677507902e-05,
987
+ "loss": 0.7919,
988
+ "step": 4050
989
+ },
990
+ {
991
+ "epoch": 1.58,
992
+ "learning_rate": 9.181670705690761e-05,
993
+ "loss": 0.7919,
994
+ "step": 4075
995
+ },
996
+ {
997
+ "epoch": 1.59,
998
+ "learning_rate": 9.080412277777413e-05,
999
+ "loss": 0.8018,
1000
+ "step": 4100
1001
+ },
1002
+ {
1003
+ "epoch": 1.6,
1004
+ "learning_rate": 8.979248855061188e-05,
1005
+ "loss": 0.7811,
1006
+ "step": 4125
1007
+ },
1008
+ {
1009
+ "epoch": 1.61,
1010
+ "learning_rate": 8.878190889020159e-05,
1011
+ "loss": 0.7919,
1012
+ "step": 4150
1013
+ },
1014
+ {
1015
+ "epoch": 1.62,
1016
+ "learning_rate": 8.777248820237376e-05,
1017
+ "loss": 0.7994,
1018
+ "step": 4175
1019
+ },
1020
+ {
1021
+ "epoch": 1.63,
1022
+ "learning_rate": 8.676433077322215e-05,
1023
+ "loss": 0.7956,
1024
+ "step": 4200
1025
+ }
1026
+ ],
1027
+ "max_steps": 7737,
1028
+ "num_train_epochs": 3,
1029
+ "total_flos": 1.8086435753964863e+19,
1030
+ "trial_name": null,
1031
+ "trial_params": null
1032
+ }
checkpoint-4200/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c24a630a19529cd8950cf57e21bd075e18eda756219dfbf6103ddefc1c70630c
3
+ size 4027
checkpoint-4300/README.md ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - load_in_8bit: False
9
+ - load_in_4bit: True
10
+ - llm_int8_threshold: 6.0
11
+ - llm_int8_skip_modules: None
12
+ - llm_int8_enable_fp32_cpu_offload: False
13
+ - llm_int8_has_fp16_weight: False
14
+ - bnb_4bit_quant_type: nf4
15
+ - bnb_4bit_use_double_quant: True
16
+ - bnb_4bit_compute_dtype: bfloat16
17
+ ### Framework versions
18
+
19
+
20
+ - PEFT 0.5.0.dev0
checkpoint-4300/adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "/workspace/webui/models/TheBloke_Llama-2-13B-fp16",
4
+ "bias": "none",
5
+ "fan_in_fan_out": null,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 16,
11
+ "lora_dropout": 0.05,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 32,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "down_proj",
18
+ "o_proj",
19
+ "v_proj",
20
+ "k_proj",
21
+ "up_proj",
22
+ "gate_proj",
23
+ "q_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM"
26
+ }
checkpoint-4300/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c41a72072ab56250ace68437ed6dfbf71629850b2bc273c736b51e3a496356f
3
+ size 500897101
checkpoint-4300/adapter_model/README.md ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - load_in_8bit: False
9
+ - load_in_4bit: True
10
+ - llm_int8_threshold: 6.0
11
+ - llm_int8_skip_modules: None
12
+ - llm_int8_enable_fp32_cpu_offload: False
13
+ - llm_int8_has_fp16_weight: False
14
+ - bnb_4bit_quant_type: nf4
15
+ - bnb_4bit_use_double_quant: True
16
+ - bnb_4bit_compute_dtype: bfloat16
17
+ ### Framework versions
18
+
19
+
20
+ - PEFT 0.5.0.dev0
checkpoint-4300/adapter_model/adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "/workspace/webui/models/TheBloke_Llama-2-13B-fp16",
4
+ "bias": "none",
5
+ "fan_in_fan_out": null,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 16,
11
+ "lora_dropout": 0.05,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 32,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "down_proj",
18
+ "o_proj",
19
+ "v_proj",
20
+ "k_proj",
21
+ "up_proj",
22
+ "gate_proj",
23
+ "q_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM"
26
+ }
checkpoint-4300/adapter_model/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c41a72072ab56250ace68437ed6dfbf71629850b2bc273c736b51e3a496356f
3
+ size 500897101
checkpoint-4300/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:497e6c81664dc7b983f91f99863172e9574fdc3b3300cd5cea395680c9252510
3
+ size 1001752701
checkpoint-4300/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:121ef05de7997793447d03c8cc4e266d9073eeaa3daedb55dc7e25140ef02dd1
3
+ size 27772
checkpoint-4300/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d24925171643f6ea34b80f875d9a1d50dcaf47c4e71d1be81cd1b39185ee9d0
3
+ size 27772
checkpoint-4300/rng_state_10.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bedaa9c487c46c3801eb08ab178fc11278a3bf7e2726ebf48e074b55b5fa7fd8
3
+ size 27789
checkpoint-4300/rng_state_11.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff2f62570667272eeab7b3fc587eff6d7cdce355a33c310dde1c1a03249034e2
3
+ size 27789
checkpoint-4300/rng_state_12.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f51d71d8b000087b6e8f1209e96ec36897cf5e5b510c84d753be0556bb5e293e
3
+ size 27789
checkpoint-4300/rng_state_13.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a254c2e04d9cb39c6423862bd7e75ed481a97371074a634c1a8d18aaea950684
3
+ size 27789
checkpoint-4300/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13a4ee228bcada53d78aca82f1f6bf2e6809aef2e31b6ed10833aaf48978be24
3
+ size 27772
checkpoint-4300/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bf390318fe3129233c66f25cd42699bad45503cc7186c232c074d1c8c9d907f
3
+ size 27772
checkpoint-4300/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:050973cfeb069667ebb795720649153bf5e63ddf73b320403aeaf0ae40fff3ac
3
+ size 27772
checkpoint-4300/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c09ee56150dad40f49dde724f23ce615070c660ccfc241f3c576fb8c20e2a33
3
+ size 27772
checkpoint-4300/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf2598e9e35dd77d2a0a105a48390f264e19bfc113eb7658cfd53e1980249715
3
+ size 27772
checkpoint-4300/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f57d640c31db45c0e291981ef91b361505e8556ee11b4a206ce881c390bbd63e
3
+ size 27772
checkpoint-4300/rng_state_8.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4966871c7c21d3db8e1f65f611d6d30f331198039d6974c575da5d3ae49c8981
3
+ size 27772
checkpoint-4300/rng_state_9.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65ea77ce5108d37f57fbb6d39e7782060aea56f1f661ba3221397f5b97a72b9d
3
+ size 27772
checkpoint-4300/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a123b18c3753f82103026ba972ea80038fecb52008cf9e45d5e8d38d62098300
3
+ size 627
checkpoint-4300/trainer_state.json ADDED
@@ -0,0 +1,1056 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.667312911981388,
5
+ "global_step": 4300,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.02,
12
+ "learning_rate": 0.0001999867761371633,
13
+ "loss": 1.0435,
14
+ "step": 50
15
+ },
16
+ {
17
+ "epoch": 0.04,
18
+ "learning_rate": 0.00019993306018843102,
19
+ "loss": 0.8918,
20
+ "step": 100
21
+ },
22
+ {
23
+ "epoch": 0.06,
24
+ "learning_rate": 0.00019983804784290833,
25
+ "loss": 0.8874,
26
+ "step": 150
27
+ },
28
+ {
29
+ "epoch": 0.08,
30
+ "learning_rate": 0.00019970177836355307,
31
+ "loss": 0.8839,
32
+ "step": 200
33
+ },
34
+ {
35
+ "epoch": 0.09,
36
+ "learning_rate": 0.00019961818913082012,
37
+ "loss": 0.8801,
38
+ "step": 225
39
+ },
40
+ {
41
+ "epoch": 0.1,
42
+ "learning_rate": 0.00019952430806244534,
43
+ "loss": 0.8753,
44
+ "step": 250
45
+ },
46
+ {
47
+ "epoch": 0.11,
48
+ "learning_rate": 0.00019942014485754635,
49
+ "loss": 0.8754,
50
+ "step": 275
51
+ },
52
+ {
53
+ "epoch": 0.12,
54
+ "learning_rate": 0.00019930571027751713,
55
+ "loss": 0.8751,
56
+ "step": 300
57
+ },
58
+ {
59
+ "epoch": 0.13,
60
+ "learning_rate": 0.0001991810161449164,
61
+ "loss": 0.8819,
62
+ "step": 325
63
+ },
64
+ {
65
+ "epoch": 0.14,
66
+ "learning_rate": 0.00019904607534224612,
67
+ "loss": 0.8744,
68
+ "step": 350
69
+ },
70
+ {
71
+ "epoch": 0.15,
72
+ "learning_rate": 0.00019890090181062063,
73
+ "loss": 0.8735,
74
+ "step": 375
75
+ },
76
+ {
77
+ "epoch": 0.16,
78
+ "learning_rate": 0.00019874551054832625,
79
+ "loss": 0.8703,
80
+ "step": 400
81
+ },
82
+ {
83
+ "epoch": 0.16,
84
+ "learning_rate": 0.00019857991760927193,
85
+ "loss": 0.8715,
86
+ "step": 425
87
+ },
88
+ {
89
+ "epoch": 0.17,
90
+ "learning_rate": 0.00019840414010133045,
91
+ "loss": 0.8714,
92
+ "step": 450
93
+ },
94
+ {
95
+ "epoch": 0.18,
96
+ "learning_rate": 0.00019821819618457114,
97
+ "loss": 0.8653,
98
+ "step": 475
99
+ },
100
+ {
101
+ "epoch": 0.19,
102
+ "learning_rate": 0.0001980221050693837,
103
+ "loss": 0.8716,
104
+ "step": 500
105
+ },
106
+ {
107
+ "epoch": 0.2,
108
+ "learning_rate": 0.00019781588701449338,
109
+ "loss": 0.8695,
110
+ "step": 525
111
+ },
112
+ {
113
+ "epoch": 0.21,
114
+ "learning_rate": 0.0001975995633248682,
115
+ "loss": 0.8746,
116
+ "step": 550
117
+ },
118
+ {
119
+ "epoch": 0.22,
120
+ "learning_rate": 0.00019737315634951762,
121
+ "loss": 0.8731,
122
+ "step": 575
123
+ },
124
+ {
125
+ "epoch": 0.23,
126
+ "learning_rate": 0.00019713668947918386,
127
+ "loss": 0.867,
128
+ "step": 600
129
+ },
130
+ {
131
+ "epoch": 0.24,
132
+ "learning_rate": 0.0001968901871439252,
133
+ "loss": 0.8706,
134
+ "step": 625
135
+ },
136
+ {
137
+ "epoch": 0.25,
138
+ "learning_rate": 0.000196633674810592,
139
+ "loss": 0.8595,
140
+ "step": 650
141
+ },
142
+ {
143
+ "epoch": 0.26,
144
+ "learning_rate": 0.0001963671789801958,
145
+ "loss": 0.8627,
146
+ "step": 675
147
+ },
148
+ {
149
+ "epoch": 0.27,
150
+ "learning_rate": 0.0001960907271851712,
151
+ "loss": 0.8607,
152
+ "step": 700
153
+ },
154
+ {
155
+ "epoch": 0.28,
156
+ "learning_rate": 0.00019580434798653173,
157
+ "loss": 0.858,
158
+ "step": 725
159
+ },
160
+ {
161
+ "epoch": 0.29,
162
+ "learning_rate": 0.00019550807097091876,
163
+ "loss": 0.8589,
164
+ "step": 750
165
+ },
166
+ {
167
+ "epoch": 0.3,
168
+ "learning_rate": 0.00019520192674754515,
169
+ "loss": 0.8561,
170
+ "step": 775
171
+ },
172
+ {
173
+ "epoch": 0.31,
174
+ "learning_rate": 0.00019488594694503264,
175
+ "loss": 0.8576,
176
+ "step": 800
177
+ },
178
+ {
179
+ "epoch": 0.32,
180
+ "learning_rate": 0.00019456016420814446,
181
+ "loss": 0.8597,
182
+ "step": 825
183
+ },
184
+ {
185
+ "epoch": 0.33,
186
+ "learning_rate": 0.00019422461219441254,
187
+ "loss": 0.862,
188
+ "step": 850
189
+ },
190
+ {
191
+ "epoch": 0.34,
192
+ "learning_rate": 0.00019387932557066035,
193
+ "loss": 0.8577,
194
+ "step": 875
195
+ },
196
+ {
197
+ "epoch": 0.35,
198
+ "learning_rate": 0.00019352434000942127,
199
+ "loss": 0.8632,
200
+ "step": 900
201
+ },
202
+ {
203
+ "epoch": 0.36,
204
+ "learning_rate": 0.00019315969218525333,
205
+ "loss": 0.8567,
206
+ "step": 925
207
+ },
208
+ {
209
+ "epoch": 0.37,
210
+ "learning_rate": 0.00019278541977095005,
211
+ "loss": 0.8501,
212
+ "step": 950
213
+ },
214
+ {
215
+ "epoch": 0.38,
216
+ "learning_rate": 0.00019240156143364844,
217
+ "loss": 0.8596,
218
+ "step": 975
219
+ },
220
+ {
221
+ "epoch": 0.39,
222
+ "learning_rate": 0.00019200815683083434,
223
+ "loss": 0.8556,
224
+ "step": 1000
225
+ },
226
+ {
227
+ "epoch": 0.39,
228
+ "eval_loss": 0.8521950244903564,
229
+ "eval_runtime": 59.8838,
230
+ "eval_samples_per_second": 12.19,
231
+ "eval_steps_per_second": 0.885,
232
+ "step": 1000
233
+ },
234
+ {
235
+ "epoch": 0.4,
236
+ "learning_rate": 0.00019160524660624505,
237
+ "loss": 0.8531,
238
+ "step": 1025
239
+ },
240
+ {
241
+ "epoch": 0.41,
242
+ "learning_rate": 0.00019119287238567045,
243
+ "loss": 0.8513,
244
+ "step": 1050
245
+ },
246
+ {
247
+ "epoch": 0.42,
248
+ "learning_rate": 0.00019077107677265253,
249
+ "loss": 0.8502,
250
+ "step": 1075
251
+ },
252
+ {
253
+ "epoch": 0.43,
254
+ "learning_rate": 0.00019033990334408384,
255
+ "loss": 0.8469,
256
+ "step": 1100
257
+ },
258
+ {
259
+ "epoch": 0.44,
260
+ "learning_rate": 0.00018989939664570545,
261
+ "loss": 0.8495,
262
+ "step": 1125
263
+ },
264
+ {
265
+ "epoch": 0.45,
266
+ "learning_rate": 0.00018944960218750484,
267
+ "loss": 0.8485,
268
+ "step": 1150
269
+ },
270
+ {
271
+ "epoch": 0.46,
272
+ "learning_rate": 0.00018899056643901404,
273
+ "loss": 0.8534,
274
+ "step": 1175
275
+ },
276
+ {
277
+ "epoch": 0.47,
278
+ "learning_rate": 0.00018852233682450893,
279
+ "loss": 0.8531,
280
+ "step": 1200
281
+ },
282
+ {
283
+ "epoch": 0.47,
284
+ "learning_rate": 0.00018804496171810948,
285
+ "loss": 0.8509,
286
+ "step": 1225
287
+ },
288
+ {
289
+ "epoch": 0.48,
290
+ "learning_rate": 0.00018755849043878222,
291
+ "loss": 0.8445,
292
+ "step": 1250
293
+ },
294
+ {
295
+ "epoch": 0.49,
296
+ "learning_rate": 0.0001870629732452449,
297
+ "loss": 0.8548,
298
+ "step": 1275
299
+ },
300
+ {
301
+ "epoch": 0.5,
302
+ "learning_rate": 0.00018655846133077417,
303
+ "loss": 0.8441,
304
+ "step": 1300
305
+ },
306
+ {
307
+ "epoch": 0.51,
308
+ "learning_rate": 0.00018604500681791656,
309
+ "loss": 0.8533,
310
+ "step": 1325
311
+ },
312
+ {
313
+ "epoch": 0.52,
314
+ "learning_rate": 0.00018552266275310373,
315
+ "loss": 0.8505,
316
+ "step": 1350
317
+ },
318
+ {
319
+ "epoch": 0.53,
320
+ "learning_rate": 0.0001849914831011719,
321
+ "loss": 0.8544,
322
+ "step": 1375
323
+ },
324
+ {
325
+ "epoch": 0.54,
326
+ "learning_rate": 0.00018445152273978668,
327
+ "loss": 0.845,
328
+ "step": 1400
329
+ },
330
+ {
331
+ "epoch": 0.55,
332
+ "learning_rate": 0.00018390283745377354,
333
+ "loss": 0.8376,
334
+ "step": 1425
335
+ },
336
+ {
337
+ "epoch": 0.56,
338
+ "learning_rate": 0.0001833454839293545,
339
+ "loss": 0.847,
340
+ "step": 1450
341
+ },
342
+ {
343
+ "epoch": 0.57,
344
+ "learning_rate": 0.00018277951974829163,
345
+ "loss": 0.8473,
346
+ "step": 1475
347
+ },
348
+ {
349
+ "epoch": 0.58,
350
+ "learning_rate": 0.0001822050033819382,
351
+ "loss": 0.8438,
352
+ "step": 1500
353
+ },
354
+ {
355
+ "epoch": 0.59,
356
+ "learning_rate": 0.00018162199418519785,
357
+ "loss": 0.8418,
358
+ "step": 1525
359
+ },
360
+ {
361
+ "epoch": 0.6,
362
+ "learning_rate": 0.00018103055239039243,
363
+ "loss": 0.842,
364
+ "step": 1550
365
+ },
366
+ {
367
+ "epoch": 0.61,
368
+ "learning_rate": 0.0001804307391010393,
369
+ "loss": 0.8435,
370
+ "step": 1575
371
+ },
372
+ {
373
+ "epoch": 0.62,
374
+ "learning_rate": 0.00017982261628553842,
375
+ "loss": 0.8349,
376
+ "step": 1600
377
+ },
378
+ {
379
+ "epoch": 0.63,
380
+ "learning_rate": 0.0001792062467707703,
381
+ "loss": 0.8483,
382
+ "step": 1625
383
+ },
384
+ {
385
+ "epoch": 0.64,
386
+ "learning_rate": 0.0001785816942356052,
387
+ "loss": 0.8387,
388
+ "step": 1650
389
+ },
390
+ {
391
+ "epoch": 0.65,
392
+ "learning_rate": 0.00017794902320432429,
393
+ "loss": 0.843,
394
+ "step": 1675
395
+ },
396
+ {
397
+ "epoch": 0.66,
398
+ "learning_rate": 0.00017730829903995333,
399
+ "loss": 0.8424,
400
+ "step": 1700
401
+ },
402
+ {
403
+ "epoch": 0.67,
404
+ "learning_rate": 0.00017665958793751006,
405
+ "loss": 0.8418,
406
+ "step": 1725
407
+ },
408
+ {
409
+ "epoch": 0.68,
410
+ "learning_rate": 0.00017600295691716522,
411
+ "loss": 0.8384,
412
+ "step": 1750
413
+ },
414
+ {
415
+ "epoch": 0.69,
416
+ "learning_rate": 0.00017533847381731856,
417
+ "loss": 0.8445,
418
+ "step": 1775
419
+ },
420
+ {
421
+ "epoch": 0.7,
422
+ "learning_rate": 0.00017466620728759033,
423
+ "loss": 0.8446,
424
+ "step": 1800
425
+ },
426
+ {
427
+ "epoch": 0.71,
428
+ "learning_rate": 0.00017398622678172878,
429
+ "loss": 0.838,
430
+ "step": 1825
431
+ },
432
+ {
433
+ "epoch": 0.72,
434
+ "learning_rate": 0.0001732986025504348,
435
+ "loss": 0.8415,
436
+ "step": 1850
437
+ },
438
+ {
439
+ "epoch": 0.73,
440
+ "learning_rate": 0.000172603405634104,
441
+ "loss": 0.8357,
442
+ "step": 1875
443
+ },
444
+ {
445
+ "epoch": 0.74,
446
+ "learning_rate": 0.00017190070785548755,
447
+ "loss": 0.8311,
448
+ "step": 1900
449
+ },
450
+ {
451
+ "epoch": 0.75,
452
+ "learning_rate": 0.0001711905818122717,
453
+ "loss": 0.8333,
454
+ "step": 1925
455
+ },
456
+ {
457
+ "epoch": 0.76,
458
+ "learning_rate": 0.0001704731008695777,
459
+ "loss": 0.8387,
460
+ "step": 1950
461
+ },
462
+ {
463
+ "epoch": 0.77,
464
+ "learning_rate": 0.0001697483391523821,
465
+ "loss": 0.8442,
466
+ "step": 1975
467
+ },
468
+ {
469
+ "epoch": 0.78,
470
+ "learning_rate": 0.00016901637153785885,
471
+ "loss": 0.8399,
472
+ "step": 2000
473
+ },
474
+ {
475
+ "epoch": 0.78,
476
+ "eval_loss": 0.8339959383010864,
477
+ "eval_runtime": 58.5829,
478
+ "eval_samples_per_second": 12.461,
479
+ "eval_steps_per_second": 0.905,
480
+ "step": 2000
481
+ },
482
+ {
483
+ "epoch": 0.79,
484
+ "learning_rate": 0.0001682772736476434,
485
+ "loss": 0.8334,
486
+ "step": 2025
487
+ },
488
+ {
489
+ "epoch": 0.79,
490
+ "learning_rate": 0.0001675311218400201,
491
+ "loss": 0.835,
492
+ "step": 2050
493
+ },
494
+ {
495
+ "epoch": 0.8,
496
+ "learning_rate": 0.00016677799320203332,
497
+ "loss": 0.8368,
498
+ "step": 2075
499
+ },
500
+ {
501
+ "epoch": 0.81,
502
+ "learning_rate": 0.00016601796554152344,
503
+ "loss": 0.8278,
504
+ "step": 2100
505
+ },
506
+ {
507
+ "epoch": 0.82,
508
+ "learning_rate": 0.00016525111737908827,
509
+ "loss": 0.8334,
510
+ "step": 2125
511
+ },
512
+ {
513
+ "epoch": 0.83,
514
+ "learning_rate": 0.00016447752793997096,
515
+ "loss": 0.8416,
516
+ "step": 2150
517
+ },
518
+ {
519
+ "epoch": 0.84,
520
+ "learning_rate": 0.00016369727714587483,
521
+ "loss": 0.8297,
522
+ "step": 2175
523
+ },
524
+ {
525
+ "epoch": 0.85,
526
+ "learning_rate": 0.0001629104456067066,
527
+ "loss": 0.8327,
528
+ "step": 2200
529
+ },
530
+ {
531
+ "epoch": 0.86,
532
+ "learning_rate": 0.00016211711461224825,
533
+ "loss": 0.8324,
534
+ "step": 2225
535
+ },
536
+ {
537
+ "epoch": 0.87,
538
+ "learning_rate": 0.0001613173661237589,
539
+ "loss": 0.8313,
540
+ "step": 2250
541
+ },
542
+ {
543
+ "epoch": 0.88,
544
+ "learning_rate": 0.0001605112827655069,
545
+ "loss": 0.8292,
546
+ "step": 2275
547
+ },
548
+ {
549
+ "epoch": 0.89,
550
+ "learning_rate": 0.0001596989478162339,
551
+ "loss": 0.8334,
552
+ "step": 2300
553
+ },
554
+ {
555
+ "epoch": 0.9,
556
+ "learning_rate": 0.00015888044520055106,
557
+ "loss": 0.8352,
558
+ "step": 2325
559
+ },
560
+ {
561
+ "epoch": 0.91,
562
+ "learning_rate": 0.00015805585948026852,
563
+ "loss": 0.823,
564
+ "step": 2350
565
+ },
566
+ {
567
+ "epoch": 0.92,
568
+ "learning_rate": 0.000157225275845659,
569
+ "loss": 0.8293,
570
+ "step": 2375
571
+ },
572
+ {
573
+ "epoch": 0.93,
574
+ "learning_rate": 0.00015638878010665672,
575
+ "loss": 0.8289,
576
+ "step": 2400
577
+ },
578
+ {
579
+ "epoch": 0.94,
580
+ "learning_rate": 0.00015554645868399205,
581
+ "loss": 0.832,
582
+ "step": 2425
583
+ },
584
+ {
585
+ "epoch": 0.95,
586
+ "learning_rate": 0.00015469839860026308,
587
+ "loss": 0.8294,
588
+ "step": 2450
589
+ },
590
+ {
591
+ "epoch": 0.96,
592
+ "learning_rate": 0.0001538446874709452,
593
+ "loss": 0.8281,
594
+ "step": 2475
595
+ },
596
+ {
597
+ "epoch": 0.97,
598
+ "learning_rate": 0.00015298541349533925,
599
+ "loss": 0.8314,
600
+ "step": 2500
601
+ },
602
+ {
603
+ "epoch": 0.98,
604
+ "learning_rate": 0.00015212066544745926,
605
+ "loss": 0.831,
606
+ "step": 2525
607
+ },
608
+ {
609
+ "epoch": 0.99,
610
+ "learning_rate": 0.00015125053266686124,
611
+ "loss": 0.8319,
612
+ "step": 2550
613
+ },
614
+ {
615
+ "epoch": 1.0,
616
+ "learning_rate": 0.00015037510504941303,
617
+ "loss": 0.8259,
618
+ "step": 2575
619
+ },
620
+ {
621
+ "epoch": 1.01,
622
+ "learning_rate": 0.00014949447303800695,
623
+ "loss": 0.8133,
624
+ "step": 2600
625
+ },
626
+ {
627
+ "epoch": 1.02,
628
+ "learning_rate": 0.00014860872761321593,
629
+ "loss": 0.8139,
630
+ "step": 2625
631
+ },
632
+ {
633
+ "epoch": 1.03,
634
+ "learning_rate": 0.00014771796028389405,
635
+ "loss": 0.804,
636
+ "step": 2650
637
+ },
638
+ {
639
+ "epoch": 1.04,
640
+ "learning_rate": 0.0001468222630777225,
641
+ "loss": 0.8011,
642
+ "step": 2675
643
+ },
644
+ {
645
+ "epoch": 1.05,
646
+ "learning_rate": 0.00014592172853170193,
647
+ "loss": 0.8037,
648
+ "step": 2700
649
+ },
650
+ {
651
+ "epoch": 1.06,
652
+ "learning_rate": 0.00014501644968259212,
653
+ "loss": 0.8063,
654
+ "step": 2725
655
+ },
656
+ {
657
+ "epoch": 1.07,
658
+ "learning_rate": 0.00014410652005730025,
659
+ "loss": 0.8155,
660
+ "step": 2750
661
+ },
662
+ {
663
+ "epoch": 1.08,
664
+ "learning_rate": 0.00014319203366321826,
665
+ "loss": 0.8066,
666
+ "step": 2775
667
+ },
668
+ {
669
+ "epoch": 1.09,
670
+ "learning_rate": 0.0001422730849785107,
671
+ "loss": 0.8091,
672
+ "step": 2800
673
+ },
674
+ {
675
+ "epoch": 1.1,
676
+ "learning_rate": 0.0001413497689423539,
677
+ "loss": 0.8067,
678
+ "step": 2825
679
+ },
680
+ {
681
+ "epoch": 1.11,
682
+ "learning_rate": 0.00014042218094512755,
683
+ "loss": 0.8046,
684
+ "step": 2850
685
+ },
686
+ {
687
+ "epoch": 1.11,
688
+ "learning_rate": 0.00013949041681855985,
689
+ "loss": 0.8053,
690
+ "step": 2875
691
+ },
692
+ {
693
+ "epoch": 1.12,
694
+ "learning_rate": 0.0001385545728258264,
695
+ "loss": 0.8075,
696
+ "step": 2900
697
+ },
698
+ {
699
+ "epoch": 1.13,
700
+ "learning_rate": 0.0001376147456516055,
701
+ "loss": 0.8015,
702
+ "step": 2925
703
+ },
704
+ {
705
+ "epoch": 1.14,
706
+ "learning_rate": 0.00013667103239208903,
707
+ "loss": 0.8016,
708
+ "step": 2950
709
+ },
710
+ {
711
+ "epoch": 1.15,
712
+ "learning_rate": 0.00013572353054495126,
713
+ "loss": 0.8029,
714
+ "step": 2975
715
+ },
716
+ {
717
+ "epoch": 1.16,
718
+ "learning_rate": 0.0001347723379992762,
719
+ "loss": 0.8017,
720
+ "step": 3000
721
+ },
722
+ {
723
+ "epoch": 1.16,
724
+ "eval_loss": 0.8229297995567322,
725
+ "eval_runtime": 59.3398,
726
+ "eval_samples_per_second": 12.302,
727
+ "eval_steps_per_second": 0.893,
728
+ "step": 3000
729
+ },
730
+ {
731
+ "epoch": 1.17,
732
+ "learning_rate": 0.0001338175530254443,
733
+ "loss": 0.8049,
734
+ "step": 3025
735
+ },
736
+ {
737
+ "epoch": 1.18,
738
+ "learning_rate": 0.00013285927426497985,
739
+ "loss": 0.8027,
740
+ "step": 3050
741
+ },
742
+ {
743
+ "epoch": 1.19,
744
+ "learning_rate": 0.00013189760072036008,
745
+ "loss": 0.8028,
746
+ "step": 3075
747
+ },
748
+ {
749
+ "epoch": 1.2,
750
+ "learning_rate": 0.0001309326317447869,
751
+ "loss": 0.8021,
752
+ "step": 3100
753
+ },
754
+ {
755
+ "epoch": 1.21,
756
+ "learning_rate": 0.00012996446703192257,
757
+ "loss": 0.8033,
758
+ "step": 3125
759
+ },
760
+ {
761
+ "epoch": 1.22,
762
+ "learning_rate": 0.00012899320660558986,
763
+ "loss": 0.8016,
764
+ "step": 3150
765
+ },
766
+ {
767
+ "epoch": 1.23,
768
+ "learning_rate": 0.00012801895080943846,
769
+ "loss": 0.7995,
770
+ "step": 3175
771
+ },
772
+ {
773
+ "epoch": 1.24,
774
+ "learning_rate": 0.0001270418002965782,
775
+ "loss": 0.799,
776
+ "step": 3200
777
+ },
778
+ {
779
+ "epoch": 1.25,
780
+ "learning_rate": 0.0001260618560191802,
781
+ "loss": 0.8002,
782
+ "step": 3225
783
+ },
784
+ {
785
+ "epoch": 1.26,
786
+ "learning_rate": 0.00012507921921804717,
787
+ "loss": 0.8068,
788
+ "step": 3250
789
+ },
790
+ {
791
+ "epoch": 1.27,
792
+ "learning_rate": 0.00012409399141215423,
793
+ "loss": 0.8041,
794
+ "step": 3275
795
+ },
796
+ {
797
+ "epoch": 1.28,
798
+ "learning_rate": 0.0001231062743881603,
799
+ "loss": 0.7999,
800
+ "step": 3300
801
+ },
802
+ {
803
+ "epoch": 1.29,
804
+ "learning_rate": 0.0001221161701898926,
805
+ "loss": 0.7995,
806
+ "step": 3325
807
+ },
808
+ {
809
+ "epoch": 1.3,
810
+ "learning_rate": 0.00012112378110780391,
811
+ "loss": 0.7959,
812
+ "step": 3350
813
+ },
814
+ {
815
+ "epoch": 1.31,
816
+ "learning_rate": 0.00012012920966840486,
817
+ "loss": 0.7999,
818
+ "step": 3375
819
+ },
820
+ {
821
+ "epoch": 1.32,
822
+ "learning_rate": 0.00011913255862367151,
823
+ "loss": 0.8016,
824
+ "step": 3400
825
+ },
826
+ {
827
+ "epoch": 1.33,
828
+ "learning_rate": 0.00011813393094042993,
829
+ "loss": 0.7944,
830
+ "step": 3425
831
+ },
832
+ {
833
+ "epoch": 1.34,
834
+ "learning_rate": 0.0001171334297897181,
835
+ "loss": 0.8026,
836
+ "step": 3450
837
+ },
838
+ {
839
+ "epoch": 1.35,
840
+ "learning_rate": 0.00011613115853612734,
841
+ "loss": 0.8004,
842
+ "step": 3475
843
+ },
844
+ {
845
+ "epoch": 1.36,
846
+ "learning_rate": 0.00011512722072712321,
847
+ "loss": 0.7992,
848
+ "step": 3500
849
+ },
850
+ {
851
+ "epoch": 1.37,
852
+ "learning_rate": 0.00011412172008234785,
853
+ "loss": 0.8004,
854
+ "step": 3525
855
+ },
856
+ {
857
+ "epoch": 1.38,
858
+ "learning_rate": 0.0001131147604829043,
859
+ "loss": 0.8009,
860
+ "step": 3550
861
+ },
862
+ {
863
+ "epoch": 1.39,
864
+ "learning_rate": 0.00011210644596062439,
865
+ "loss": 0.7993,
866
+ "step": 3575
867
+ },
868
+ {
869
+ "epoch": 1.4,
870
+ "learning_rate": 0.00011109688068732081,
871
+ "loss": 0.7965,
872
+ "step": 3600
873
+ },
874
+ {
875
+ "epoch": 1.41,
876
+ "learning_rate": 0.00011008616896402482,
877
+ "loss": 0.7991,
878
+ "step": 3625
879
+ },
880
+ {
881
+ "epoch": 1.42,
882
+ "learning_rate": 0.00010907441521021072,
883
+ "loss": 0.8026,
884
+ "step": 3650
885
+ },
886
+ {
887
+ "epoch": 1.42,
888
+ "learning_rate": 0.00010806172395300789,
889
+ "loss": 0.7941,
890
+ "step": 3675
891
+ },
892
+ {
893
+ "epoch": 1.43,
894
+ "learning_rate": 0.00010704819981640186,
895
+ "loss": 0.7989,
896
+ "step": 3700
897
+ },
898
+ {
899
+ "epoch": 1.44,
900
+ "learning_rate": 0.00010603394751042522,
901
+ "loss": 0.7981,
902
+ "step": 3725
903
+ },
904
+ {
905
+ "epoch": 1.45,
906
+ "learning_rate": 0.00010501907182033979,
907
+ "loss": 0.7985,
908
+ "step": 3750
909
+ },
910
+ {
911
+ "epoch": 1.46,
912
+ "learning_rate": 0.000104003677595811,
913
+ "loss": 0.7921,
914
+ "step": 3775
915
+ },
916
+ {
917
+ "epoch": 1.47,
918
+ "learning_rate": 0.00010298786974007555,
919
+ "loss": 0.8012,
920
+ "step": 3800
921
+ },
922
+ {
923
+ "epoch": 1.48,
924
+ "learning_rate": 0.00010197175319910343,
925
+ "loss": 0.7906,
926
+ "step": 3825
927
+ },
928
+ {
929
+ "epoch": 1.49,
930
+ "learning_rate": 0.00010095543295075593,
931
+ "loss": 0.7928,
932
+ "step": 3850
933
+ },
934
+ {
935
+ "epoch": 1.5,
936
+ "learning_rate": 9.993901399393979e-05,
937
+ "loss": 0.8018,
938
+ "step": 3875
939
+ },
940
+ {
941
+ "epoch": 1.51,
942
+ "learning_rate": 9.892260133775968e-05,
943
+ "loss": 0.7991,
944
+ "step": 3900
945
+ },
946
+ {
947
+ "epoch": 1.52,
948
+ "learning_rate": 9.79062999906693e-05,
949
+ "loss": 0.795,
950
+ "step": 3925
951
+ },
952
+ {
953
+ "epoch": 1.53,
954
+ "learning_rate": 9.68902149496227e-05,
955
+ "loss": 0.7977,
956
+ "step": 3950
957
+ },
958
+ {
959
+ "epoch": 1.54,
960
+ "learning_rate": 9.587445118922674e-05,
961
+ "loss": 0.8013,
962
+ "step": 3975
963
+ },
964
+ {
965
+ "epoch": 1.55,
966
+ "learning_rate": 9.485911365089589e-05,
967
+ "loss": 0.7978,
968
+ "step": 4000
969
+ },
970
+ {
971
+ "epoch": 1.55,
972
+ "eval_loss": 0.8142631649971008,
973
+ "eval_runtime": 59.4108,
974
+ "eval_samples_per_second": 12.287,
975
+ "eval_steps_per_second": 0.892,
976
+ "step": 4000
977
+ },
978
+ {
979
+ "epoch": 1.56,
980
+ "learning_rate": 9.384430723201036e-05,
981
+ "loss": 0.7912,
982
+ "step": 4025
983
+ },
984
+ {
985
+ "epoch": 1.57,
986
+ "learning_rate": 9.283013677507902e-05,
987
+ "loss": 0.7919,
988
+ "step": 4050
989
+ },
990
+ {
991
+ "epoch": 1.58,
992
+ "learning_rate": 9.181670705690761e-05,
993
+ "loss": 0.7919,
994
+ "step": 4075
995
+ },
996
+ {
997
+ "epoch": 1.59,
998
+ "learning_rate": 9.080412277777413e-05,
999
+ "loss": 0.8018,
1000
+ "step": 4100
1001
+ },
1002
+ {
1003
+ "epoch": 1.6,
1004
+ "learning_rate": 8.979248855061188e-05,
1005
+ "loss": 0.7811,
1006
+ "step": 4125
1007
+ },
1008
+ {
1009
+ "epoch": 1.61,
1010
+ "learning_rate": 8.878190889020159e-05,
1011
+ "loss": 0.7919,
1012
+ "step": 4150
1013
+ },
1014
+ {
1015
+ "epoch": 1.62,
1016
+ "learning_rate": 8.777248820237376e-05,
1017
+ "loss": 0.7994,
1018
+ "step": 4175
1019
+ },
1020
+ {
1021
+ "epoch": 1.63,
1022
+ "learning_rate": 8.676433077322215e-05,
1023
+ "loss": 0.7956,
1024
+ "step": 4200
1025
+ },
1026
+ {
1027
+ "epoch": 1.64,
1028
+ "learning_rate": 8.575754075832973e-05,
1029
+ "loss": 0.7968,
1030
+ "step": 4225
1031
+ },
1032
+ {
1033
+ "epoch": 1.65,
1034
+ "learning_rate": 8.475222217200801e-05,
1035
+ "loss": 0.7905,
1036
+ "step": 4250
1037
+ },
1038
+ {
1039
+ "epoch": 1.66,
1040
+ "learning_rate": 8.374847887655112e-05,
1041
+ "loss": 0.7889,
1042
+ "step": 4275
1043
+ },
1044
+ {
1045
+ "epoch": 1.67,
1046
+ "learning_rate": 8.274641457150543e-05,
1047
+ "loss": 0.7988,
1048
+ "step": 4300
1049
+ }
1050
+ ],
1051
+ "max_steps": 7737,
1052
+ "num_train_epochs": 3,
1053
+ "total_flos": 1.8516859681104724e+19,
1054
+ "trial_name": null,
1055
+ "trial_params": null
1056
+ }
checkpoint-4300/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c24a630a19529cd8950cf57e21bd075e18eda756219dfbf6103ddefc1c70630c
3
+ size 4027
checkpoint-4400/README.md ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - load_in_8bit: False
9
+ - load_in_4bit: True
10
+ - llm_int8_threshold: 6.0
11
+ - llm_int8_skip_modules: None
12
+ - llm_int8_enable_fp32_cpu_offload: False
13
+ - llm_int8_has_fp16_weight: False
14
+ - bnb_4bit_quant_type: nf4
15
+ - bnb_4bit_use_double_quant: True
16
+ - bnb_4bit_compute_dtype: bfloat16
17
+ ### Framework versions
18
+
19
+
20
+ - PEFT 0.5.0.dev0
checkpoint-4400/adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "/workspace/webui/models/TheBloke_Llama-2-13B-fp16",
4
+ "bias": "none",
5
+ "fan_in_fan_out": null,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 16,
11
+ "lora_dropout": 0.05,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 32,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "gate_proj",
18
+ "v_proj",
19
+ "k_proj",
20
+ "q_proj",
21
+ "up_proj",
22
+ "o_proj",
23
+ "down_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM"
26
+ }