Jingmei commited on
Commit
115837d
1 Parent(s): 8f55737

Training in progress, step 8

Browse files
adapter_config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/global/cfs/cdirs/m3623/jmyang/Pandmeic_LLM/model/Llama-3.2-1B",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.1,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 128,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "q_proj",
24
+ "v_proj",
25
+ "o_proj",
26
+ "k_proj"
27
+ ],
28
+ "task_type": "CAUSAL_LM",
29
+ "use_dora": false,
30
+ "use_rslora": false
31
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abee0fe1bc0d4da04a8b25f8460113db0916b3e530e32e752b1cc63df39b1044
3
+ size 109069232
trainer_peft.log ADDED
@@ -0,0 +1,180 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-11-08 13:38 - Start running the training script
2
+ 2024-11-08 13:38 - Start running the training script
3
+ 2024-11-08 13:38 - Start running the training script
4
+ 2024-11-08 13:38 - Start running the training script
5
+ 2024-11-08 13:38 - Cuda check
6
+ 2024-11-08 13:38 - Cuda check
7
+ 2024-11-08 13:38 - Cuda check
8
+ 2024-11-08 13:38 - Cuda check
9
+ 2024-11-08 13:38 - False
10
+ 2024-11-08 13:38 - False
11
+ 2024-11-08 13:38 - False
12
+ 2024-11-08 13:38 - False
13
+ 2024-11-08 13:38 - 0
14
+ 2024-11-08 13:38 - Configue Model and tokenizer
15
+ 2024-11-08 13:38 - 0
16
+ 2024-11-08 13:38 - Configue Model and tokenizer
17
+ 2024-11-08 13:38 - 0
18
+ 2024-11-08 13:38 - Configue Model and tokenizer
19
+ 2024-11-08 13:38 - 0
20
+ 2024-11-08 13:38 - Configue Model and tokenizer
21
+ 2024-11-08 13:38 - Complete configuring Model
22
+ 2024-11-08 13:38 - Configue tokenizer
23
+ 2024-11-08 13:38 - Complete configuring Model
24
+ 2024-11-08 13:38 - Complete configuring Model
25
+ 2024-11-08 13:38 - Configue tokenizer
26
+ 2024-11-08 13:38 - Configue tokenizer
27
+ 2024-11-08 13:38 - Complete configuring Model
28
+ 2024-11-08 13:38 - Configue tokenizer
29
+ 2024-11-08 13:38 - Complete configuring tokenizer
30
+ 2024-11-08 13:38 - Load dataset
31
+ 2024-11-08 13:38 - Complete configuring tokenizer
32
+ 2024-11-08 13:38 - Load dataset
33
+ 2024-11-08 13:38 - Complete configuring tokenizer
34
+ 2024-11-08 13:38 - Load dataset
35
+ 2024-11-08 13:38 - Complete configuring tokenizer
36
+ 2024-11-08 13:38 - Load dataset
37
+ 2024-11-08 13:42 - Start running the training script
38
+ 2024-11-08 13:42 - Start running the training script
39
+ 2024-11-08 13:42 - Start running the training script
40
+ 2024-11-08 13:42 - Start running the training script
41
+ 2024-11-08 13:42 - Cuda check
42
+ 2024-11-08 13:42 - Cuda check
43
+ 2024-11-08 13:42 - Cuda check
44
+ 2024-11-08 13:42 - Cuda check
45
+ 2024-11-08 13:42 - False
46
+ 2024-11-08 13:42 - False
47
+ 2024-11-08 13:42 - False
48
+ 2024-11-08 13:42 - False
49
+ 2024-11-08 13:42 - 0
50
+ 2024-11-08 13:42 - 0
51
+ 2024-11-08 13:42 - Configue Model and tokenizer
52
+ 2024-11-08 13:42 - Configue Model and tokenizer
53
+ 2024-11-08 13:42 - 0
54
+ 2024-11-08 13:42 - Configue Model and tokenizer
55
+ 2024-11-08 13:42 - 0
56
+ 2024-11-08 13:42 - Configue Model and tokenizer
57
+ 2024-11-08 13:42 - Complete configuring Model
58
+ 2024-11-08 13:42 - Complete configuring Model
59
+ 2024-11-08 13:42 - Configue tokenizer
60
+ 2024-11-08 13:42 - Configue tokenizer
61
+ 2024-11-08 13:42 - Complete configuring Model
62
+ 2024-11-08 13:42 - Configue tokenizer
63
+ 2024-11-08 13:42 - Complete configuring Model
64
+ 2024-11-08 13:42 - Configue tokenizer
65
+ 2024-11-08 13:42 - Complete configuring tokenizer
66
+ 2024-11-08 13:42 - Load dataset
67
+ 2024-11-08 13:42 - Complete configuring tokenizer
68
+ 2024-11-08 13:42 - Load dataset
69
+ 2024-11-08 13:42 - Complete configuring tokenizer
70
+ 2024-11-08 13:42 - Load dataset
71
+ 2024-11-08 13:42 - Complete configuring tokenizer
72
+ 2024-11-08 13:42 - Load dataset
73
+ 2024-11-08 13:46 - Start running the training script
74
+ 2024-11-08 13:46 - Cuda check
75
+ 2024-11-08 13:46 - True
76
+ 2024-11-08 13:46 - 4
77
+ 2024-11-08 13:46 - Configue Model and tokenizer
78
+ 2024-11-08 13:46 - Start running the training script
79
+ 2024-11-08 13:46 - Start running the training script
80
+ 2024-11-08 13:46 - Cuda check
81
+ 2024-11-08 13:46 - Cuda check
82
+ 2024-11-08 13:46 - True
83
+ 2024-11-08 13:46 - 4
84
+ 2024-11-08 13:46 - True
85
+ 2024-11-08 13:46 - Configue Model and tokenizer
86
+ 2024-11-08 13:46 - 4
87
+ 2024-11-08 13:46 - Configue Model and tokenizer
88
+ 2024-11-08 13:46 - Start running the training script
89
+ 2024-11-08 13:46 - Cuda check
90
+ 2024-11-08 13:46 - True
91
+ 2024-11-08 13:46 - 4
92
+ 2024-11-08 13:46 - Configue Model and tokenizer
93
+ 2024-11-08 13:46 - Complete configuring Model
94
+ 2024-11-08 13:46 - Configue tokenizer
95
+ 2024-11-08 13:46 - Complete configuring Model
96
+ 2024-11-08 13:46 - Configue tokenizer
97
+ 2024-11-08 13:46 - Complete configuring Model
98
+ 2024-11-08 13:46 - Configue tokenizer
99
+ 2024-11-08 13:46 - Complete configuring Model
100
+ 2024-11-08 13:46 - Configue tokenizer
101
+ 2024-11-08 13:46 - Complete configuring tokenizer
102
+ 2024-11-08 13:46 - Load dataset
103
+ 2024-11-08 13:46 - Complete configuring tokenizer
104
+ 2024-11-08 13:46 - Load dataset
105
+ 2024-11-08 13:46 - Complete configuring tokenizer
106
+ 2024-11-08 13:46 - Load dataset
107
+ 2024-11-08 13:46 - Complete configuring tokenizer
108
+ 2024-11-08 13:46 - Load dataset
109
+ 2024-11-08 13:55 - Start running the training script
110
+ 2024-11-08 13:55 - Cuda check
111
+ 2024-11-08 13:55 - True
112
+ 2024-11-08 13:55 - 4
113
+ 2024-11-08 13:55 - Configue Model and tokenizer
114
+ 2024-11-08 13:55 - Start running the training script
115
+ 2024-11-08 13:55 - Start running the training script
116
+ 2024-11-08 13:55 - Cuda check
117
+ 2024-11-08 13:55 - Cuda check
118
+ 2024-11-08 13:55 - True
119
+ 2024-11-08 13:55 - True
120
+ 2024-11-08 13:55 - 4
121
+ 2024-11-08 13:55 - 4
122
+ 2024-11-08 13:55 - Configue Model and tokenizer
123
+ 2024-11-08 13:55 - Configue Model and tokenizer
124
+ 2024-11-08 13:55 - Start running the training script
125
+ 2024-11-08 13:55 - Cuda check
126
+ 2024-11-08 13:55 - True
127
+ 2024-11-08 13:55 - 4
128
+ 2024-11-08 13:55 - Configue Model and tokenizer
129
+ 2024-11-08 13:55 - Complete configuring Model
130
+ 2024-11-08 13:55 - Configue tokenizer
131
+ 2024-11-08 13:55 - Complete configuring Model
132
+ 2024-11-08 13:55 - Configue tokenizer
133
+ 2024-11-08 13:55 - Complete configuring Model
134
+ 2024-11-08 13:55 - Configue tokenizer
135
+ 2024-11-08 13:55 - Complete configuring Model
136
+ 2024-11-08 13:55 - Configue tokenizer
137
+ 2024-11-08 13:55 - Complete configuring tokenizer
138
+ 2024-11-08 13:55 - Load dataset
139
+ 2024-11-08 13:55 - Complete configuring tokenizer
140
+ 2024-11-08 13:55 - Load dataset
141
+ 2024-11-08 13:55 - Complete configuring tokenizer
142
+ 2024-11-08 13:55 - Load dataset
143
+ 2024-11-08 13:55 - Complete configuring tokenizer
144
+ 2024-11-08 13:55 - Load dataset
145
+ 2024-11-08 15:13 - Start running the training script
146
+ 2024-11-08 15:13 - Start running the training script
147
+ 2024-11-08 15:13 - Cuda check
148
+ 2024-11-08 15:13 - True
149
+ 2024-11-08 15:13 - 4
150
+ 2024-11-08 15:13 - Configue Model and tokenizer
151
+ 2024-11-08 15:13 - Cuda check
152
+ 2024-11-08 15:13 - True
153
+ 2024-11-08 15:13 - 4
154
+ 2024-11-08 15:13 - Configue Model and tokenizer
155
+ 2024-11-08 15:13 - Start running the training script
156
+ 2024-11-08 15:13 - Cuda check
157
+ 2024-11-08 15:13 - True
158
+ 2024-11-08 15:13 - 4
159
+ 2024-11-08 15:13 - Configue Model and tokenizer
160
+ 2024-11-08 15:14 - Start running the training script
161
+ 2024-11-08 15:14 - Cuda check
162
+ 2024-11-08 15:14 - True
163
+ 2024-11-08 15:14 - 4
164
+ 2024-11-08 15:14 - Configue Model and tokenizer
165
+ 2024-11-08 15:14 - Complete configuring Model
166
+ 2024-11-08 15:14 - Configue tokenizer
167
+ 2024-11-08 15:14 - Complete configuring Model
168
+ 2024-11-08 15:14 - Complete configuring Model
169
+ 2024-11-08 15:14 - Complete configuring Model
170
+ 2024-11-08 15:14 - Configue tokenizer
171
+ 2024-11-08 15:14 - Configue tokenizer
172
+ 2024-11-08 15:14 - Configue tokenizer
173
+ 2024-11-08 15:14 - Complete configuring tokenizer
174
+ 2024-11-08 15:14 - Load dataset
175
+ 2024-11-08 15:14 - Complete configuring tokenizer
176
+ 2024-11-08 15:14 - Load dataset
177
+ 2024-11-08 15:14 - Complete configuring tokenizer
178
+ 2024-11-08 15:14 - Load dataset
179
+ 2024-11-08 15:14 - Complete configuring tokenizer
180
+ 2024-11-08 15:14 - Load dataset
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99c3ca59dade593d2fa51fb2a46df601928df74fa6ec15129b09eefbf6d544ca
3
+ size 5368