wissamantoun commited on
Commit
6a6117d
1 Parent(s): ee907fb

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,277 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: fr
3
+ license: mit
4
+ tags:
5
+ - roberta
6
+ - text-classification
7
+ base_model: almanach/camembertv2-base
8
+ datasets:
9
+ - FLUE-PAWS-X
10
+ metrics:
11
+ - accuracy
12
+ pipeline_tag: text-classification
13
+ library_name: transformers
14
+ model-index:
15
+ - name: almanach/camembertv2-base-pawsx
16
+ results:
17
+ - task:
18
+ type: text-classification
19
+ name: Paraphrase Identification
20
+ dataset:
21
+ type: flue-paws-x
22
+ name: FLUE-PAWS-X
23
+ metrics:
24
+ - name: accuracy
25
+ type: accuracy
26
+ value: 0.92254
27
+ verified: false
28
+ ---
29
+
30
+ # Model Card for almanach/camembertv2-base-pawsx
31
+
32
+ almanach/camembertv2-base-pawsx is a roberta model for text classification. It is trained on the FLUE-PAWS-X dataset for the task of Paraphrase Identification. The model achieves an accuracy of 0.92254 on the FLUE-PAWS-X dataset.
33
+
34
+ The model is part of the almanach/camembertv2-base family of model finetunes.
35
+
36
+ ## Model Details
37
+
38
+ ### Model Description
39
+
40
+ - **Developed by:** Wissam Antoun (Phd Student at Almanach, Inria-Paris)
41
+ - **Model type:** roberta
42
+ - **Language(s) (NLP):** French
43
+ - **License:** MIT
44
+ - **Finetuned from model [optional]:** almanach/camembertv2-base
45
+
46
+ ### Model Sources [optional]
47
+
48
+ <!-- Provide the basic links for the model. -->
49
+
50
+ - **Repository:** https://github.com/WissamAntoun/camemberta
51
+ - **Paper:** https://arxiv.org/abs/2411.08868
52
+
53
+ ## Uses
54
+
55
+ The model can be used for text classification tasks in French for Paraphrase Identification.
56
+
57
+ ## Bias, Risks, and Limitations
58
+
59
+ The model may exhibit biases based on the training data. The model may not generalize well to other datasets or tasks. The model may also have limitations in terms of the data it was trained on.
60
+
61
+
62
+ ## How to Get Started with the Model
63
+
64
+ Use the code below to get started with the model.
65
+
66
+ ```python
67
+ from transformers import AutoTokenizer, AutoModelForSequenceClassification, pipeline
68
+
69
+ model = AutoModelForSequenceClassification.from_pretrained("almanach/camembertv2-base-pawsx")
70
+ tokenizer = AutoTokenizer.from_pretrained("almanach/camembertv2-base-pawsx")
71
+
72
+ classifier = pipeline("text-classification", model=model, tokenizer=tokenizer)
73
+
74
+ classifier({
75
+ "text": "Le livre est très intéressant et j'ai appris beaucoup de choses.",
76
+ "text_pair": "Le livre est très ennuyeux et je n'ai rien appris.",
77
+ })
78
+ ```
79
+
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ The model is trained on the FLUE-PAWS-X dataset.
86
+
87
+ - Dataset Name: FLUE-PAWS-X
88
+ - Dataset Size:
89
+ - Train: 49399
90
+ - Dev: 1988
91
+ - Test: 2000
92
+
93
+
94
+ ### Training Procedure
95
+
96
+ Model trained with the run_classification.py script from the huggingface repository.
97
+
98
+
99
+
100
+ #### Training Hyperparameters
101
+
102
+ ```yml
103
+ accelerator_config: '{''split_batches'': False, ''dispatch_batches'': None, ''even_batches'':
104
+ True, ''use_seedable_sampler'': True, ''non_blocking'': False, ''gradient_accumulation_kwargs'':
105
+ None}'
106
+ adafactor: false
107
+ adam_beta1: 0.9
108
+ adam_beta2: 0.999
109
+ adam_epsilon: 1.0e-08
110
+ auto_find_batch_size: false
111
+ base_model: camembertv2
112
+ base_model_name: camembertv2-base-bf16-p2-17000
113
+ batch_eval_metrics: false
114
+ bf16: false
115
+ bf16_full_eval: false
116
+ data_seed: 1.0
117
+ dataloader_drop_last: false
118
+ dataloader_num_workers: 0
119
+ dataloader_persistent_workers: false
120
+ dataloader_pin_memory: true
121
+ dataloader_prefetch_factor: .nan
122
+ ddp_backend: .nan
123
+ ddp_broadcast_buffers: .nan
124
+ ddp_bucket_cap_mb: .nan
125
+ ddp_find_unused_parameters: .nan
126
+ ddp_timeout: 1800
127
+ debug: '[]'
128
+ deepspeed: .nan
129
+ disable_tqdm: false
130
+ dispatch_batches: .nan
131
+ do_eval: true
132
+ do_predict: false
133
+ do_train: true
134
+ epoch: 5.999028340080971
135
+ eval_accumulation_steps: 4
136
+ eval_accuracy: 0.9225352112676056
137
+ eval_delay: 0
138
+ eval_do_concat_batches: true
139
+ eval_loss: 0.3642682433128357
140
+ eval_on_start: false
141
+ eval_runtime: 4.0364
142
+ eval_samples: 1988
143
+ eval_samples_per_second: 492.519
144
+ eval_steps: .nan
145
+ eval_steps_per_second: 61.689
146
+ eval_strategy: epoch
147
+ eval_use_gather_object: false
148
+ evaluation_strategy: epoch
149
+ fp16: false
150
+ fp16_backend: auto
151
+ fp16_full_eval: false
152
+ fp16_opt_level: O1
153
+ fsdp: '[]'
154
+ fsdp_config: '{''min_num_params'': 0, ''xla'': False, ''xla_fsdp_v2'': False, ''xla_fsdp_grad_ckpt'':
155
+ False}'
156
+ fsdp_min_num_params: 0
157
+ fsdp_transformer_layer_cls_to_wrap: .nan
158
+ full_determinism: false
159
+ gradient_accumulation_steps: 2
160
+ gradient_checkpointing: false
161
+ gradient_checkpointing_kwargs: .nan
162
+ greater_is_better: true
163
+ group_by_length: false
164
+ half_precision_backend: auto
165
+ hub_always_push: false
166
+ hub_model_id: .nan
167
+ hub_private_repo: false
168
+ hub_strategy: every_save
169
+ hub_token: <HUB_TOKEN>
170
+ ignore_data_skip: false
171
+ include_inputs_for_metrics: false
172
+ include_num_input_tokens_seen: false
173
+ include_tokens_per_second: false
174
+ jit_mode_eval: false
175
+ label_names: .nan
176
+ label_smoothing_factor: 0.0
177
+ learning_rate: 3.0e-05
178
+ length_column_name: length
179
+ load_best_model_at_end: true
180
+ local_rank: 0
181
+ log_level: debug
182
+ log_level_replica: warning
183
+ log_on_each_node: true
184
+ logging_dir: /scratch/camembertv2/runs/results/flue-PAWS-X/camembertv2-base-bf16-p2-17000/max_seq_length-148-gradient_accumulation_steps-2-precision-fp32-learning_rate-3e-05-epochs-6-lr_scheduler-linear-warmup_steps-0/SEED-1/logs
185
+ logging_first_step: false
186
+ logging_nan_inf_filter: true
187
+ logging_steps: 100
188
+ logging_strategy: steps
189
+ lr_scheduler_kwargs: '{}'
190
+ lr_scheduler_type: linear
191
+ max_grad_norm: 1.0
192
+ max_steps: -1
193
+ metric_for_best_model: accuracy
194
+ mp_parameters: .nan
195
+ name: camembertv2/runs/results/flue-PAWS-X/camembertv2-base-bf16-p2-17000/max_seq_length-148-gradient_accumulation_steps-2-precision-fp32-learning_rate-3e-05-epochs-6-lr_scheduler-linear-warmup_steps-0
196
+ neftune_noise_alpha: .nan
197
+ no_cuda: false
198
+ num_train_epochs: 6.0
199
+ optim: adamw_torch
200
+ optim_args: .nan
201
+ optim_target_modules: .nan
202
+ output_dir: /scratch/camembertv2/runs/results/flue-PAWS-X/camembertv2-base-bf16-p2-17000/max_seq_length-148-gradient_accumulation_steps-2-precision-fp32-learning_rate-3e-05-epochs-6-lr_scheduler-linear-warmup_steps-0/SEED-1
203
+ overwrite_output_dir: false
204
+ past_index: -1
205
+ per_device_eval_batch_size: 8
206
+ per_device_train_batch_size: 8
207
+ per_gpu_eval_batch_size: .nan
208
+ per_gpu_train_batch_size: .nan
209
+ prediction_loss_only: false
210
+ push_to_hub: false
211
+ push_to_hub_model_id: .nan
212
+ push_to_hub_organization: .nan
213
+ push_to_hub_token: <PUSH_TO_HUB_TOKEN>
214
+ ray_scope: last
215
+ remove_unused_columns: true
216
+ report_to: '[''tensorboard'']'
217
+ restore_callback_states_from_checkpoint: false
218
+ resume_from_checkpoint: .nan
219
+ run_name: /scratch/camembertv2/runs/results/flue-PAWS-X/camembertv2-base-bf16-p2-17000/max_seq_length-148-gradient_accumulation_steps-2-precision-fp32-learning_rate-3e-05-epochs-6-lr_scheduler-linear-warmup_steps-0/SEED-1
220
+ save_on_each_node: false
221
+ save_only_model: false
222
+ save_safetensors: true
223
+ save_steps: 500
224
+ save_strategy: epoch
225
+ save_total_limit: .nan
226
+ seed: 1
227
+ skip_memory_metrics: true
228
+ split_batches: .nan
229
+ tf32: .nan
230
+ torch_compile: true
231
+ torch_compile_backend: inductor
232
+ torch_compile_mode: .nan
233
+ torch_empty_cache_steps: .nan
234
+ torchdynamo: .nan
235
+ total_flos: 1.33712370278538e+16
236
+ tpu_metrics_debug: false
237
+ tpu_num_cores: .nan
238
+ train_loss: 0.1708474308300605
239
+ train_runtime: 2225.7449
240
+ train_samples: 49399
241
+ train_samples_per_second: 133.166
242
+ train_steps_per_second: 8.322
243
+ use_cpu: false
244
+ use_ipex: false
245
+ use_legacy_prediction_loop: false
246
+ use_mps_device: false
247
+ warmup_ratio: 0.0
248
+ warmup_steps: 0
249
+ weight_decay: 0.0
250
+
251
+ ```
252
+
253
+ #### Results
254
+
255
+ **Accuracy:** 0.92254
256
+
257
+ ## Technical Specifications
258
+
259
+ ### Model Architecture and Objective
260
+
261
+ roberta for sequence classification.
262
+
263
+ ## Citation
264
+
265
+ **BibTeX:**
266
+
267
+ ```bibtex
268
+ @misc{antoun2024camembert20smarterfrench,
269
+ title={CamemBERT 2.0: A Smarter French Language Model Aged to Perfection},
270
+ author={Wissam Antoun and Francis Kulumba and Rian Touchent and Éric de la Clergerie and Benoît Sagot and Djamé Seddah},
271
+ year={2024},
272
+ eprint={2411.08868},
273
+ archivePrefix={arXiv},
274
+ primaryClass={cs.CL},
275
+ url={https://arxiv.org/abs/2411.08868},
276
+ }
277
+ ```
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 5.999028340080971,
3
+ "eval_accuracy": 0.9225352112676056,
4
+ "eval_loss": 0.3642682433128357,
5
+ "eval_runtime": 4.0364,
6
+ "eval_samples": 1988,
7
+ "eval_samples_per_second": 492.519,
8
+ "eval_steps_per_second": 61.689,
9
+ "total_flos": 1.33712370278538e+16,
10
+ "train_loss": 0.17084743083006051,
11
+ "train_runtime": 2225.7449,
12
+ "train_samples": 49399,
13
+ "train_samples_per_second": 133.166,
14
+ "train_steps_per_second": 8.322
15
+ }
config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/scratch/camembertv2/runs/models/camembertv2-base-bf16/post/ckpt-p2-17000/pt/",
3
+ "architectures": [
4
+ "RobertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 1,
8
+ "classifier_dropout": null,
9
+ "embedding_size": 768,
10
+ "eos_token_id": 2,
11
+ "finetuning_task": "paws-x",
12
+ "hidden_act": "gelu",
13
+ "hidden_dropout_prob": 0.1,
14
+ "hidden_size": 768,
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 3072,
17
+ "label2id": {
18
+ "0": 0,
19
+ "1": 1
20
+ },
21
+ "layer_norm_eps": 1e-07,
22
+ "max_position_embeddings": 1025,
23
+ "model_name": "camembertv2-base-bf16",
24
+ "model_type": "roberta",
25
+ "num_attention_heads": 12,
26
+ "num_hidden_layers": 12,
27
+ "pad_token_id": 0,
28
+ "position_biased_input": true,
29
+ "position_embedding_type": "absolute",
30
+ "problem_type": "single_label_classification",
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.44.2",
33
+ "type_vocab_size": 1,
34
+ "use_cache": true,
35
+ "vocab_size": 32768
36
+ }
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 5.999028340080971,
3
+ "eval_accuracy": 0.9225352112676056,
4
+ "eval_loss": 0.3642682433128357,
5
+ "eval_runtime": 4.0364,
6
+ "eval_samples": 1988,
7
+ "eval_samples_per_second": 492.519,
8
+ "eval_steps_per_second": 61.689
9
+ }
logs/events.out.tfevents.1724562251.nefgpu37.208119.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5f8873a076f9c7f265aec46c23352dc68e453781821e61077b5d1b13306299b
3
+ size 47105
logs/events.out.tfevents.1724564481.nefgpu37.208119.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ac298647e77f3b658757e4a58ca4b09c2fcab2641cf7cc6f71f10557887dc1b
3
+ size 369
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04c1c28eaa3530ea771566c034e0c01ae5390d2b3ba3edf4d147aae1253b1f67
3
+ size 446431832
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "[CLS]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "[SEP]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "[MASK]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "[PAD]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "[SEP]",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "[UNK]",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "[PAD]",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "[CLS]",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "[SEP]",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "[UNK]",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "4": {
37
+ "content": "[MASK]",
38
+ "lstrip": false,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
+ "bos_token": "[CLS]",
46
+ "clean_up_tokenization_spaces": true,
47
+ "cls_token": "[CLS]",
48
+ "eos_token": "[SEP]",
49
+ "errors": "replace",
50
+ "mask_token": "[MASK]",
51
+ "model_max_length": 1000000000000000019884624838656,
52
+ "pad_token": "[PAD]",
53
+ "sep_token": "[SEP]",
54
+ "tokenizer_class": "RobertaTokenizer",
55
+ "trim_offsets": true,
56
+ "unk_token": "[UNK]"
57
+ }
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 5.999028340080971,
3
+ "total_flos": 1.33712370278538e+16,
4
+ "train_loss": 0.17084743083006051,
5
+ "train_runtime": 2225.7449,
6
+ "train_samples": 49399,
7
+ "train_samples_per_second": 133.166,
8
+ "train_steps_per_second": 8.322
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,1391 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.9225352112676056,
3
+ "best_model_checkpoint": "/scratch/camembertv2/runs/results/flue-PAWS-X/camembertv2-base-bf16-p2-17000/max_seq_length-148-gradient_accumulation_steps-2-precision-fp32-learning_rate-3e-05-epochs-6-lr_scheduler-linear-warmup_steps-0/SEED-1/checkpoint-15437",
4
+ "epoch": 5.999028340080971,
5
+ "eval_steps": 500,
6
+ "global_step": 18522,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.032388663967611336,
13
+ "grad_norm": 1.3617459535598755,
14
+ "learning_rate": 2.983803045027535e-05,
15
+ "loss": 0.6917,
16
+ "step": 100
17
+ },
18
+ {
19
+ "epoch": 0.06477732793522267,
20
+ "grad_norm": 1.1722190380096436,
21
+ "learning_rate": 2.96760609005507e-05,
22
+ "loss": 0.6863,
23
+ "step": 200
24
+ },
25
+ {
26
+ "epoch": 0.09716599190283401,
27
+ "grad_norm": 5.13236665725708,
28
+ "learning_rate": 2.9514091350826045e-05,
29
+ "loss": 0.5495,
30
+ "step": 300
31
+ },
32
+ {
33
+ "epoch": 0.12955465587044535,
34
+ "grad_norm": 11.99015998840332,
35
+ "learning_rate": 2.9352121801101394e-05,
36
+ "loss": 0.4849,
37
+ "step": 400
38
+ },
39
+ {
40
+ "epoch": 0.16194331983805668,
41
+ "grad_norm": 7.261778831481934,
42
+ "learning_rate": 2.9190152251376742e-05,
43
+ "loss": 0.4368,
44
+ "step": 500
45
+ },
46
+ {
47
+ "epoch": 0.19433198380566802,
48
+ "grad_norm": 8.82419490814209,
49
+ "learning_rate": 2.9028182701652093e-05,
50
+ "loss": 0.4046,
51
+ "step": 600
52
+ },
53
+ {
54
+ "epoch": 0.22672064777327935,
55
+ "grad_norm": 7.460156440734863,
56
+ "learning_rate": 2.8866213151927438e-05,
57
+ "loss": 0.3746,
58
+ "step": 700
59
+ },
60
+ {
61
+ "epoch": 0.2591093117408907,
62
+ "grad_norm": 17.467742919921875,
63
+ "learning_rate": 2.8704243602202786e-05,
64
+ "loss": 0.3493,
65
+ "step": 800
66
+ },
67
+ {
68
+ "epoch": 0.291497975708502,
69
+ "grad_norm": 16.750350952148438,
70
+ "learning_rate": 2.8542274052478135e-05,
71
+ "loss": 0.3583,
72
+ "step": 900
73
+ },
74
+ {
75
+ "epoch": 0.32388663967611336,
76
+ "grad_norm": 8.11487102508545,
77
+ "learning_rate": 2.8380304502753486e-05,
78
+ "loss": 0.3469,
79
+ "step": 1000
80
+ },
81
+ {
82
+ "epoch": 0.3562753036437247,
83
+ "grad_norm": 13.264874458312988,
84
+ "learning_rate": 2.821833495302883e-05,
85
+ "loss": 0.3359,
86
+ "step": 1100
87
+ },
88
+ {
89
+ "epoch": 0.38866396761133604,
90
+ "grad_norm": 38.216304779052734,
91
+ "learning_rate": 2.805636540330418e-05,
92
+ "loss": 0.32,
93
+ "step": 1200
94
+ },
95
+ {
96
+ "epoch": 0.42105263157894735,
97
+ "grad_norm": 3.897590398788452,
98
+ "learning_rate": 2.7894395853579527e-05,
99
+ "loss": 0.3134,
100
+ "step": 1300
101
+ },
102
+ {
103
+ "epoch": 0.4534412955465587,
104
+ "grad_norm": 19.601438522338867,
105
+ "learning_rate": 2.773242630385488e-05,
106
+ "loss": 0.3247,
107
+ "step": 1400
108
+ },
109
+ {
110
+ "epoch": 0.48582995951417,
111
+ "grad_norm": 6.229040145874023,
112
+ "learning_rate": 2.7570456754130224e-05,
113
+ "loss": 0.3283,
114
+ "step": 1500
115
+ },
116
+ {
117
+ "epoch": 0.5182186234817814,
118
+ "grad_norm": 8.433351516723633,
119
+ "learning_rate": 2.7408487204405572e-05,
120
+ "loss": 0.2851,
121
+ "step": 1600
122
+ },
123
+ {
124
+ "epoch": 0.5506072874493927,
125
+ "grad_norm": 29.696481704711914,
126
+ "learning_rate": 2.724651765468092e-05,
127
+ "loss": 0.3067,
128
+ "step": 1700
129
+ },
130
+ {
131
+ "epoch": 0.582995951417004,
132
+ "grad_norm": 9.64140796661377,
133
+ "learning_rate": 2.7084548104956272e-05,
134
+ "loss": 0.3072,
135
+ "step": 1800
136
+ },
137
+ {
138
+ "epoch": 0.6153846153846154,
139
+ "grad_norm": 2.7816696166992188,
140
+ "learning_rate": 2.6922578555231617e-05,
141
+ "loss": 0.2838,
142
+ "step": 1900
143
+ },
144
+ {
145
+ "epoch": 0.6477732793522267,
146
+ "grad_norm": 11.457731246948242,
147
+ "learning_rate": 2.6760609005506965e-05,
148
+ "loss": 0.2692,
149
+ "step": 2000
150
+ },
151
+ {
152
+ "epoch": 0.680161943319838,
153
+ "grad_norm": 10.894031524658203,
154
+ "learning_rate": 2.6598639455782313e-05,
155
+ "loss": 0.2969,
156
+ "step": 2100
157
+ },
158
+ {
159
+ "epoch": 0.7125506072874493,
160
+ "grad_norm": 8.756068229675293,
161
+ "learning_rate": 2.6436669906057665e-05,
162
+ "loss": 0.2982,
163
+ "step": 2200
164
+ },
165
+ {
166
+ "epoch": 0.7449392712550608,
167
+ "grad_norm": 6.366727352142334,
168
+ "learning_rate": 2.627470035633301e-05,
169
+ "loss": 0.2743,
170
+ "step": 2300
171
+ },
172
+ {
173
+ "epoch": 0.7773279352226721,
174
+ "grad_norm": 3.6022021770477295,
175
+ "learning_rate": 2.6112730806608358e-05,
176
+ "loss": 0.2497,
177
+ "step": 2400
178
+ },
179
+ {
180
+ "epoch": 0.8097165991902834,
181
+ "grad_norm": 9.339791297912598,
182
+ "learning_rate": 2.5950761256883706e-05,
183
+ "loss": 0.2617,
184
+ "step": 2500
185
+ },
186
+ {
187
+ "epoch": 0.8421052631578947,
188
+ "grad_norm": 0.46193593740463257,
189
+ "learning_rate": 2.5788791707159058e-05,
190
+ "loss": 0.2591,
191
+ "step": 2600
192
+ },
193
+ {
194
+ "epoch": 0.8744939271255061,
195
+ "grad_norm": 8.35823917388916,
196
+ "learning_rate": 2.5626822157434402e-05,
197
+ "loss": 0.2619,
198
+ "step": 2700
199
+ },
200
+ {
201
+ "epoch": 0.9068825910931174,
202
+ "grad_norm": 11.017343521118164,
203
+ "learning_rate": 2.546485260770975e-05,
204
+ "loss": 0.2571,
205
+ "step": 2800
206
+ },
207
+ {
208
+ "epoch": 0.9392712550607287,
209
+ "grad_norm": 12.998709678649902,
210
+ "learning_rate": 2.53028830579851e-05,
211
+ "loss": 0.2922,
212
+ "step": 2900
213
+ },
214
+ {
215
+ "epoch": 0.97165991902834,
216
+ "grad_norm": 10.282382011413574,
217
+ "learning_rate": 2.514091350826045e-05,
218
+ "loss": 0.2421,
219
+ "step": 3000
220
+ },
221
+ {
222
+ "epoch": 0.9998380566801619,
223
+ "eval_accuracy": 0.8958752515090543,
224
+ "eval_loss": 0.3331478238105774,
225
+ "eval_runtime": 4.4897,
226
+ "eval_samples_per_second": 442.795,
227
+ "eval_steps_per_second": 55.461,
228
+ "step": 3087
229
+ },
230
+ {
231
+ "epoch": 1.0040485829959513,
232
+ "grad_norm": 1.8964722156524658,
233
+ "learning_rate": 2.4978943958535795e-05,
234
+ "loss": 0.2629,
235
+ "step": 3100
236
+ },
237
+ {
238
+ "epoch": 1.0364372469635628,
239
+ "grad_norm": 20.548351287841797,
240
+ "learning_rate": 2.4816974408811143e-05,
241
+ "loss": 0.251,
242
+ "step": 3200
243
+ },
244
+ {
245
+ "epoch": 1.0688259109311742,
246
+ "grad_norm": 9.91510009765625,
247
+ "learning_rate": 2.465500485908649e-05,
248
+ "loss": 0.205,
249
+ "step": 3300
250
+ },
251
+ {
252
+ "epoch": 1.1012145748987854,
253
+ "grad_norm": 15.848530769348145,
254
+ "learning_rate": 2.4493035309361843e-05,
255
+ "loss": 0.1904,
256
+ "step": 3400
257
+ },
258
+ {
259
+ "epoch": 1.1336032388663968,
260
+ "grad_norm": 2.94976544380188,
261
+ "learning_rate": 2.4331065759637188e-05,
262
+ "loss": 0.2101,
263
+ "step": 3500
264
+ },
265
+ {
266
+ "epoch": 1.165991902834008,
267
+ "grad_norm": 2.3269777297973633,
268
+ "learning_rate": 2.4169096209912536e-05,
269
+ "loss": 0.1924,
270
+ "step": 3600
271
+ },
272
+ {
273
+ "epoch": 1.1983805668016194,
274
+ "grad_norm": 9.416550636291504,
275
+ "learning_rate": 2.4007126660187884e-05,
276
+ "loss": 0.2243,
277
+ "step": 3700
278
+ },
279
+ {
280
+ "epoch": 1.2307692307692308,
281
+ "grad_norm": 18.740257263183594,
282
+ "learning_rate": 2.3845157110463236e-05,
283
+ "loss": 0.204,
284
+ "step": 3800
285
+ },
286
+ {
287
+ "epoch": 1.263157894736842,
288
+ "grad_norm": 11.511466026306152,
289
+ "learning_rate": 2.368318756073858e-05,
290
+ "loss": 0.2211,
291
+ "step": 3900
292
+ },
293
+ {
294
+ "epoch": 1.2955465587044535,
295
+ "grad_norm": 48.617271423339844,
296
+ "learning_rate": 2.352121801101393e-05,
297
+ "loss": 0.2006,
298
+ "step": 4000
299
+ },
300
+ {
301
+ "epoch": 1.3279352226720649,
302
+ "grad_norm": 1.9806264638900757,
303
+ "learning_rate": 2.3359248461289277e-05,
304
+ "loss": 0.2331,
305
+ "step": 4100
306
+ },
307
+ {
308
+ "epoch": 1.360323886639676,
309
+ "grad_norm": 11.272978782653809,
310
+ "learning_rate": 2.319727891156463e-05,
311
+ "loss": 0.228,
312
+ "step": 4200
313
+ },
314
+ {
315
+ "epoch": 1.3927125506072875,
316
+ "grad_norm": 13.344253540039062,
317
+ "learning_rate": 2.3035309361839974e-05,
318
+ "loss": 0.2166,
319
+ "step": 4300
320
+ },
321
+ {
322
+ "epoch": 1.425101214574899,
323
+ "grad_norm": 6.339032173156738,
324
+ "learning_rate": 2.2873339812115322e-05,
325
+ "loss": 0.1978,
326
+ "step": 4400
327
+ },
328
+ {
329
+ "epoch": 1.45748987854251,
330
+ "grad_norm": 12.384753227233887,
331
+ "learning_rate": 2.271137026239067e-05,
332
+ "loss": 0.2137,
333
+ "step": 4500
334
+ },
335
+ {
336
+ "epoch": 1.4898785425101215,
337
+ "grad_norm": 2.5586955547332764,
338
+ "learning_rate": 2.254940071266602e-05,
339
+ "loss": 0.1965,
340
+ "step": 4600
341
+ },
342
+ {
343
+ "epoch": 1.522267206477733,
344
+ "grad_norm": 30.300275802612305,
345
+ "learning_rate": 2.2387431162941366e-05,
346
+ "loss": 0.2013,
347
+ "step": 4700
348
+ },
349
+ {
350
+ "epoch": 1.5546558704453441,
351
+ "grad_norm": 0.9059699773788452,
352
+ "learning_rate": 2.2225461613216715e-05,
353
+ "loss": 0.215,
354
+ "step": 4800
355
+ },
356
+ {
357
+ "epoch": 1.5870445344129553,
358
+ "grad_norm": 24.580528259277344,
359
+ "learning_rate": 2.2063492063492063e-05,
360
+ "loss": 0.2025,
361
+ "step": 4900
362
+ },
363
+ {
364
+ "epoch": 1.6194331983805668,
365
+ "grad_norm": 46.1649169921875,
366
+ "learning_rate": 2.1901522513767414e-05,
367
+ "loss": 0.1775,
368
+ "step": 5000
369
+ },
370
+ {
371
+ "epoch": 1.6518218623481782,
372
+ "grad_norm": 9.169315338134766,
373
+ "learning_rate": 2.173955296404276e-05,
374
+ "loss": 0.2034,
375
+ "step": 5100
376
+ },
377
+ {
378
+ "epoch": 1.6842105263157894,
379
+ "grad_norm": 22.808425903320312,
380
+ "learning_rate": 2.1577583414318108e-05,
381
+ "loss": 0.2338,
382
+ "step": 5200
383
+ },
384
+ {
385
+ "epoch": 1.7165991902834008,
386
+ "grad_norm": 0.972823977470398,
387
+ "learning_rate": 2.1415613864593456e-05,
388
+ "loss": 0.2079,
389
+ "step": 5300
390
+ },
391
+ {
392
+ "epoch": 1.7489878542510122,
393
+ "grad_norm": 12.798443794250488,
394
+ "learning_rate": 2.1253644314868807e-05,
395
+ "loss": 0.2155,
396
+ "step": 5400
397
+ },
398
+ {
399
+ "epoch": 1.7813765182186234,
400
+ "grad_norm": 14.221925735473633,
401
+ "learning_rate": 2.1091674765144152e-05,
402
+ "loss": 0.2016,
403
+ "step": 5500
404
+ },
405
+ {
406
+ "epoch": 1.8137651821862348,
407
+ "grad_norm": 16.415285110473633,
408
+ "learning_rate": 2.09297052154195e-05,
409
+ "loss": 0.187,
410
+ "step": 5600
411
+ },
412
+ {
413
+ "epoch": 1.8461538461538463,
414
+ "grad_norm": 3.618945598602295,
415
+ "learning_rate": 2.0767735665694852e-05,
416
+ "loss": 0.2213,
417
+ "step": 5700
418
+ },
419
+ {
420
+ "epoch": 1.8785425101214575,
421
+ "grad_norm": 39.05452346801758,
422
+ "learning_rate": 2.06057661159702e-05,
423
+ "loss": 0.1785,
424
+ "step": 5800
425
+ },
426
+ {
427
+ "epoch": 1.9109311740890689,
428
+ "grad_norm": 19.066164016723633,
429
+ "learning_rate": 2.0443796566245545e-05,
430
+ "loss": 0.2173,
431
+ "step": 5900
432
+ },
433
+ {
434
+ "epoch": 1.9433198380566803,
435
+ "grad_norm": 48.79111862182617,
436
+ "learning_rate": 2.0281827016520893e-05,
437
+ "loss": 0.1846,
438
+ "step": 6000
439
+ },
440
+ {
441
+ "epoch": 1.9757085020242915,
442
+ "grad_norm": 0.30500343441963196,
443
+ "learning_rate": 2.0119857466796245e-05,
444
+ "loss": 0.2054,
445
+ "step": 6100
446
+ },
447
+ {
448
+ "epoch": 2.0,
449
+ "eval_accuracy": 0.9094567404426559,
450
+ "eval_loss": 0.30133694410324097,
451
+ "eval_runtime": 4.0681,
452
+ "eval_samples_per_second": 488.676,
453
+ "eval_steps_per_second": 61.207,
454
+ "step": 6175
455
+ },
456
+ {
457
+ "epoch": 2.0080971659919027,
458
+ "grad_norm": 5.170982837677002,
459
+ "learning_rate": 1.9957887917071593e-05,
460
+ "loss": 0.1975,
461
+ "step": 6200
462
+ },
463
+ {
464
+ "epoch": 2.0404858299595143,
465
+ "grad_norm": 17.76828956604004,
466
+ "learning_rate": 1.9795918367346938e-05,
467
+ "loss": 0.1424,
468
+ "step": 6300
469
+ },
470
+ {
471
+ "epoch": 2.0728744939271255,
472
+ "grad_norm": 0.5401751399040222,
473
+ "learning_rate": 1.9633948817622286e-05,
474
+ "loss": 0.1564,
475
+ "step": 6400
476
+ },
477
+ {
478
+ "epoch": 2.1052631578947367,
479
+ "grad_norm": 0.1927350014448166,
480
+ "learning_rate": 1.9471979267897638e-05,
481
+ "loss": 0.1559,
482
+ "step": 6500
483
+ },
484
+ {
485
+ "epoch": 2.1376518218623484,
486
+ "grad_norm": 6.872354984283447,
487
+ "learning_rate": 1.9310009718172986e-05,
488
+ "loss": 0.1577,
489
+ "step": 6600
490
+ },
491
+ {
492
+ "epoch": 2.1700404858299596,
493
+ "grad_norm": 0.9195345640182495,
494
+ "learning_rate": 1.914804016844833e-05,
495
+ "loss": 0.1655,
496
+ "step": 6700
497
+ },
498
+ {
499
+ "epoch": 2.2024291497975708,
500
+ "grad_norm": 0.5888263583183289,
501
+ "learning_rate": 1.898607061872368e-05,
502
+ "loss": 0.1404,
503
+ "step": 6800
504
+ },
505
+ {
506
+ "epoch": 2.234817813765182,
507
+ "grad_norm": 20.80263900756836,
508
+ "learning_rate": 1.882410106899903e-05,
509
+ "loss": 0.1828,
510
+ "step": 6900
511
+ },
512
+ {
513
+ "epoch": 2.2672064777327936,
514
+ "grad_norm": 11.595118522644043,
515
+ "learning_rate": 1.866213151927438e-05,
516
+ "loss": 0.1768,
517
+ "step": 7000
518
+ },
519
+ {
520
+ "epoch": 2.299595141700405,
521
+ "grad_norm": 4.5267333984375,
522
+ "learning_rate": 1.8500161969549723e-05,
523
+ "loss": 0.1546,
524
+ "step": 7100
525
+ },
526
+ {
527
+ "epoch": 2.331983805668016,
528
+ "grad_norm": 0.5598276257514954,
529
+ "learning_rate": 1.833819241982507e-05,
530
+ "loss": 0.1605,
531
+ "step": 7200
532
+ },
533
+ {
534
+ "epoch": 2.3643724696356276,
535
+ "grad_norm": 1.0213172435760498,
536
+ "learning_rate": 1.8176222870100423e-05,
537
+ "loss": 0.1715,
538
+ "step": 7300
539
+ },
540
+ {
541
+ "epoch": 2.396761133603239,
542
+ "grad_norm": 4.324398040771484,
543
+ "learning_rate": 1.801425332037577e-05,
544
+ "loss": 0.1597,
545
+ "step": 7400
546
+ },
547
+ {
548
+ "epoch": 2.42914979757085,
549
+ "grad_norm": 0.6324844360351562,
550
+ "learning_rate": 1.7852283770651116e-05,
551
+ "loss": 0.1659,
552
+ "step": 7500
553
+ },
554
+ {
555
+ "epoch": 2.4615384615384617,
556
+ "grad_norm": 0.5160787105560303,
557
+ "learning_rate": 1.7690314220926464e-05,
558
+ "loss": 0.1799,
559
+ "step": 7600
560
+ },
561
+ {
562
+ "epoch": 2.493927125506073,
563
+ "grad_norm": 0.26999789476394653,
564
+ "learning_rate": 1.7528344671201816e-05,
565
+ "loss": 0.1347,
566
+ "step": 7700
567
+ },
568
+ {
569
+ "epoch": 2.526315789473684,
570
+ "grad_norm": 4.179189682006836,
571
+ "learning_rate": 1.7366375121477164e-05,
572
+ "loss": 0.1737,
573
+ "step": 7800
574
+ },
575
+ {
576
+ "epoch": 2.5587044534412957,
577
+ "grad_norm": 14.371667861938477,
578
+ "learning_rate": 1.720440557175251e-05,
579
+ "loss": 0.1825,
580
+ "step": 7900
581
+ },
582
+ {
583
+ "epoch": 2.591093117408907,
584
+ "grad_norm": 0.13881312310695648,
585
+ "learning_rate": 1.7042436022027857e-05,
586
+ "loss": 0.1757,
587
+ "step": 8000
588
+ },
589
+ {
590
+ "epoch": 2.623481781376518,
591
+ "grad_norm": 10.17100715637207,
592
+ "learning_rate": 1.688046647230321e-05,
593
+ "loss": 0.1501,
594
+ "step": 8100
595
+ },
596
+ {
597
+ "epoch": 2.6558704453441297,
598
+ "grad_norm": 43.74755859375,
599
+ "learning_rate": 1.6718496922578557e-05,
600
+ "loss": 0.1845,
601
+ "step": 8200
602
+ },
603
+ {
604
+ "epoch": 2.688259109311741,
605
+ "grad_norm": 17.117637634277344,
606
+ "learning_rate": 1.6556527372853902e-05,
607
+ "loss": 0.1745,
608
+ "step": 8300
609
+ },
610
+ {
611
+ "epoch": 2.720647773279352,
612
+ "grad_norm": 31.1845645904541,
613
+ "learning_rate": 1.639455782312925e-05,
614
+ "loss": 0.1664,
615
+ "step": 8400
616
+ },
617
+ {
618
+ "epoch": 2.753036437246964,
619
+ "grad_norm": 9.80735969543457,
620
+ "learning_rate": 1.6232588273404602e-05,
621
+ "loss": 0.1632,
622
+ "step": 8500
623
+ },
624
+ {
625
+ "epoch": 2.785425101214575,
626
+ "grad_norm": 9.704376220703125,
627
+ "learning_rate": 1.607061872367995e-05,
628
+ "loss": 0.1443,
629
+ "step": 8600
630
+ },
631
+ {
632
+ "epoch": 2.817813765182186,
633
+ "grad_norm": 4.445658206939697,
634
+ "learning_rate": 1.5908649173955295e-05,
635
+ "loss": 0.1775,
636
+ "step": 8700
637
+ },
638
+ {
639
+ "epoch": 2.850202429149798,
640
+ "grad_norm": 1.7219343185424805,
641
+ "learning_rate": 1.5746679624230643e-05,
642
+ "loss": 0.1581,
643
+ "step": 8800
644
+ },
645
+ {
646
+ "epoch": 2.882591093117409,
647
+ "grad_norm": 8.307001113891602,
648
+ "learning_rate": 1.5584710074505995e-05,
649
+ "loss": 0.1536,
650
+ "step": 8900
651
+ },
652
+ {
653
+ "epoch": 2.91497975708502,
654
+ "grad_norm": 9.363420486450195,
655
+ "learning_rate": 1.5422740524781343e-05,
656
+ "loss": 0.1862,
657
+ "step": 9000
658
+ },
659
+ {
660
+ "epoch": 2.9473684210526314,
661
+ "grad_norm": 20.355863571166992,
662
+ "learning_rate": 1.5260770975056688e-05,
663
+ "loss": 0.1679,
664
+ "step": 9100
665
+ },
666
+ {
667
+ "epoch": 2.979757085020243,
668
+ "grad_norm": 16.128374099731445,
669
+ "learning_rate": 1.5098801425332037e-05,
670
+ "loss": 0.1513,
671
+ "step": 9200
672
+ },
673
+ {
674
+ "epoch": 2.999838056680162,
675
+ "eval_accuracy": 0.9054325955734407,
676
+ "eval_loss": 0.40321749448776245,
677
+ "eval_runtime": 4.0611,
678
+ "eval_samples_per_second": 489.52,
679
+ "eval_steps_per_second": 61.313,
680
+ "step": 9262
681
+ },
682
+ {
683
+ "epoch": 3.0121457489878543,
684
+ "grad_norm": 2.3643200397491455,
685
+ "learning_rate": 1.4936831875607386e-05,
686
+ "loss": 0.1294,
687
+ "step": 9300
688
+ },
689
+ {
690
+ "epoch": 3.0445344129554655,
691
+ "grad_norm": 0.2633157968521118,
692
+ "learning_rate": 1.4774862325882734e-05,
693
+ "loss": 0.1198,
694
+ "step": 9400
695
+ },
696
+ {
697
+ "epoch": 3.076923076923077,
698
+ "grad_norm": 116.33301544189453,
699
+ "learning_rate": 1.4612892776158082e-05,
700
+ "loss": 0.1199,
701
+ "step": 9500
702
+ },
703
+ {
704
+ "epoch": 3.1093117408906883,
705
+ "grad_norm": 110.17882537841797,
706
+ "learning_rate": 1.445092322643343e-05,
707
+ "loss": 0.1101,
708
+ "step": 9600
709
+ },
710
+ {
711
+ "epoch": 3.1417004048582995,
712
+ "grad_norm": 66.73577117919922,
713
+ "learning_rate": 1.4288953676708779e-05,
714
+ "loss": 0.1634,
715
+ "step": 9700
716
+ },
717
+ {
718
+ "epoch": 3.174089068825911,
719
+ "grad_norm": 0.10323189944028854,
720
+ "learning_rate": 1.4126984126984127e-05,
721
+ "loss": 0.1326,
722
+ "step": 9800
723
+ },
724
+ {
725
+ "epoch": 3.2064777327935223,
726
+ "grad_norm": 0.39017683267593384,
727
+ "learning_rate": 1.3965014577259475e-05,
728
+ "loss": 0.1305,
729
+ "step": 9900
730
+ },
731
+ {
732
+ "epoch": 3.2388663967611335,
733
+ "grad_norm": 13.028061866760254,
734
+ "learning_rate": 1.3803045027534823e-05,
735
+ "loss": 0.1254,
736
+ "step": 10000
737
+ },
738
+ {
739
+ "epoch": 3.2712550607287447,
740
+ "grad_norm": 3.545095205307007,
741
+ "learning_rate": 1.3641075477810171e-05,
742
+ "loss": 0.1227,
743
+ "step": 10100
744
+ },
745
+ {
746
+ "epoch": 3.3036437246963564,
747
+ "grad_norm": 0.2068515121936798,
748
+ "learning_rate": 1.347910592808552e-05,
749
+ "loss": 0.1107,
750
+ "step": 10200
751
+ },
752
+ {
753
+ "epoch": 3.3360323886639676,
754
+ "grad_norm": 0.7598180174827576,
755
+ "learning_rate": 1.3317136378360868e-05,
756
+ "loss": 0.1431,
757
+ "step": 10300
758
+ },
759
+ {
760
+ "epoch": 3.3684210526315788,
761
+ "grad_norm": 0.3546907603740692,
762
+ "learning_rate": 1.3155166828636216e-05,
763
+ "loss": 0.128,
764
+ "step": 10400
765
+ },
766
+ {
767
+ "epoch": 3.4008097165991904,
768
+ "grad_norm": 1.108842372894287,
769
+ "learning_rate": 1.2993197278911564e-05,
770
+ "loss": 0.1247,
771
+ "step": 10500
772
+ },
773
+ {
774
+ "epoch": 3.4331983805668016,
775
+ "grad_norm": 31.11785316467285,
776
+ "learning_rate": 1.2831227729186914e-05,
777
+ "loss": 0.1293,
778
+ "step": 10600
779
+ },
780
+ {
781
+ "epoch": 3.465587044534413,
782
+ "grad_norm": 0.4772971272468567,
783
+ "learning_rate": 1.266925817946226e-05,
784
+ "loss": 0.136,
785
+ "step": 10700
786
+ },
787
+ {
788
+ "epoch": 3.4979757085020244,
789
+ "grad_norm": 3.384209632873535,
790
+ "learning_rate": 1.250728862973761e-05,
791
+ "loss": 0.1197,
792
+ "step": 10800
793
+ },
794
+ {
795
+ "epoch": 3.5303643724696356,
796
+ "grad_norm": 13.939515113830566,
797
+ "learning_rate": 1.2345319080012957e-05,
798
+ "loss": 0.1414,
799
+ "step": 10900
800
+ },
801
+ {
802
+ "epoch": 3.562753036437247,
803
+ "grad_norm": 0.4108864963054657,
804
+ "learning_rate": 1.2183349530288307e-05,
805
+ "loss": 0.107,
806
+ "step": 11000
807
+ },
808
+ {
809
+ "epoch": 3.5951417004048585,
810
+ "grad_norm": 12.29635238647461,
811
+ "learning_rate": 1.2021379980563655e-05,
812
+ "loss": 0.1357,
813
+ "step": 11100
814
+ },
815
+ {
816
+ "epoch": 3.6275303643724697,
817
+ "grad_norm": 49.79674530029297,
818
+ "learning_rate": 1.1859410430839003e-05,
819
+ "loss": 0.1333,
820
+ "step": 11200
821
+ },
822
+ {
823
+ "epoch": 3.659919028340081,
824
+ "grad_norm": 0.1119026318192482,
825
+ "learning_rate": 1.1697440881114352e-05,
826
+ "loss": 0.1311,
827
+ "step": 11300
828
+ },
829
+ {
830
+ "epoch": 3.6923076923076925,
831
+ "grad_norm": 0.2087603062391281,
832
+ "learning_rate": 1.15354713313897e-05,
833
+ "loss": 0.145,
834
+ "step": 11400
835
+ },
836
+ {
837
+ "epoch": 3.7246963562753037,
838
+ "grad_norm": 0.3955753743648529,
839
+ "learning_rate": 1.1373501781665048e-05,
840
+ "loss": 0.1225,
841
+ "step": 11500
842
+ },
843
+ {
844
+ "epoch": 3.757085020242915,
845
+ "grad_norm": 0.05121416971087456,
846
+ "learning_rate": 1.1211532231940396e-05,
847
+ "loss": 0.0975,
848
+ "step": 11600
849
+ },
850
+ {
851
+ "epoch": 3.7894736842105265,
852
+ "grad_norm": 21.439380645751953,
853
+ "learning_rate": 1.1049562682215744e-05,
854
+ "loss": 0.1398,
855
+ "step": 11700
856
+ },
857
+ {
858
+ "epoch": 3.8218623481781377,
859
+ "grad_norm": 1.1940436363220215,
860
+ "learning_rate": 1.0887593132491093e-05,
861
+ "loss": 0.1231,
862
+ "step": 11800
863
+ },
864
+ {
865
+ "epoch": 3.854251012145749,
866
+ "grad_norm": 12.916104316711426,
867
+ "learning_rate": 1.072562358276644e-05,
868
+ "loss": 0.1392,
869
+ "step": 11900
870
+ },
871
+ {
872
+ "epoch": 3.8866396761133606,
873
+ "grad_norm": 0.13800889253616333,
874
+ "learning_rate": 1.0563654033041789e-05,
875
+ "loss": 0.142,
876
+ "step": 12000
877
+ },
878
+ {
879
+ "epoch": 3.919028340080972,
880
+ "grad_norm": 0.10773531347513199,
881
+ "learning_rate": 1.0401684483317137e-05,
882
+ "loss": 0.1444,
883
+ "step": 12100
884
+ },
885
+ {
886
+ "epoch": 3.951417004048583,
887
+ "grad_norm": 4.032077312469482,
888
+ "learning_rate": 1.0239714933592485e-05,
889
+ "loss": 0.127,
890
+ "step": 12200
891
+ },
892
+ {
893
+ "epoch": 3.983805668016194,
894
+ "grad_norm": 0.21410343050956726,
895
+ "learning_rate": 1.0077745383867834e-05,
896
+ "loss": 0.1475,
897
+ "step": 12300
898
+ },
899
+ {
900
+ "epoch": 4.0,
901
+ "eval_accuracy": 0.9079476861167002,
902
+ "eval_loss": 0.35845422744750977,
903
+ "eval_runtime": 4.0822,
904
+ "eval_samples_per_second": 486.989,
905
+ "eval_steps_per_second": 60.996,
906
+ "step": 12350
907
+ },
908
+ {
909
+ "epoch": 4.016194331983805,
910
+ "grad_norm": 108.23912048339844,
911
+ "learning_rate": 9.915775834143182e-06,
912
+ "loss": 0.1174,
913
+ "step": 12400
914
+ },
915
+ {
916
+ "epoch": 4.048582995951417,
917
+ "grad_norm": 0.086638443171978,
918
+ "learning_rate": 9.75380628441853e-06,
919
+ "loss": 0.1094,
920
+ "step": 12500
921
+ },
922
+ {
923
+ "epoch": 4.080971659919029,
924
+ "grad_norm": 31.371986389160156,
925
+ "learning_rate": 9.591836734693878e-06,
926
+ "loss": 0.1086,
927
+ "step": 12600
928
+ },
929
+ {
930
+ "epoch": 4.113360323886639,
931
+ "grad_norm": 0.3369753658771515,
932
+ "learning_rate": 9.429867184969226e-06,
933
+ "loss": 0.0997,
934
+ "step": 12700
935
+ },
936
+ {
937
+ "epoch": 4.145748987854251,
938
+ "grad_norm": 4.937185287475586,
939
+ "learning_rate": 9.267897635244575e-06,
940
+ "loss": 0.1187,
941
+ "step": 12800
942
+ },
943
+ {
944
+ "epoch": 4.178137651821863,
945
+ "grad_norm": 0.08810489624738693,
946
+ "learning_rate": 9.105928085519923e-06,
947
+ "loss": 0.0714,
948
+ "step": 12900
949
+ },
950
+ {
951
+ "epoch": 4.2105263157894735,
952
+ "grad_norm": 2.339200496673584,
953
+ "learning_rate": 8.943958535795271e-06,
954
+ "loss": 0.1018,
955
+ "step": 13000
956
+ },
957
+ {
958
+ "epoch": 4.242914979757085,
959
+ "grad_norm": 0.8731828331947327,
960
+ "learning_rate": 8.78198898607062e-06,
961
+ "loss": 0.0783,
962
+ "step": 13100
963
+ },
964
+ {
965
+ "epoch": 4.275303643724697,
966
+ "grad_norm": 0.03349796682596207,
967
+ "learning_rate": 8.620019436345967e-06,
968
+ "loss": 0.076,
969
+ "step": 13200
970
+ },
971
+ {
972
+ "epoch": 4.3076923076923075,
973
+ "grad_norm": 0.3485426902770996,
974
+ "learning_rate": 8.458049886621316e-06,
975
+ "loss": 0.0891,
976
+ "step": 13300
977
+ },
978
+ {
979
+ "epoch": 4.340080971659919,
980
+ "grad_norm": 0.18010343611240387,
981
+ "learning_rate": 8.296080336896664e-06,
982
+ "loss": 0.1094,
983
+ "step": 13400
984
+ },
985
+ {
986
+ "epoch": 4.372469635627531,
987
+ "grad_norm": 0.1857542097568512,
988
+ "learning_rate": 8.134110787172012e-06,
989
+ "loss": 0.0976,
990
+ "step": 13500
991
+ },
992
+ {
993
+ "epoch": 4.4048582995951415,
994
+ "grad_norm": 0.5226219296455383,
995
+ "learning_rate": 7.97214123744736e-06,
996
+ "loss": 0.1342,
997
+ "step": 13600
998
+ },
999
+ {
1000
+ "epoch": 4.437246963562753,
1001
+ "grad_norm": 0.23230472207069397,
1002
+ "learning_rate": 7.810171687722709e-06,
1003
+ "loss": 0.0997,
1004
+ "step": 13700
1005
+ },
1006
+ {
1007
+ "epoch": 4.469635627530364,
1008
+ "grad_norm": 0.15814034640789032,
1009
+ "learning_rate": 7.648202137998057e-06,
1010
+ "loss": 0.0924,
1011
+ "step": 13800
1012
+ },
1013
+ {
1014
+ "epoch": 4.502024291497976,
1015
+ "grad_norm": 6.45848274230957,
1016
+ "learning_rate": 7.486232588273405e-06,
1017
+ "loss": 0.0867,
1018
+ "step": 13900
1019
+ },
1020
+ {
1021
+ "epoch": 4.534412955465587,
1022
+ "grad_norm": 0.31769734621047974,
1023
+ "learning_rate": 7.324263038548753e-06,
1024
+ "loss": 0.1186,
1025
+ "step": 14000
1026
+ },
1027
+ {
1028
+ "epoch": 4.566801619433198,
1029
+ "grad_norm": 5.155035495758057,
1030
+ "learning_rate": 7.162293488824101e-06,
1031
+ "loss": 0.0912,
1032
+ "step": 14100
1033
+ },
1034
+ {
1035
+ "epoch": 4.59919028340081,
1036
+ "grad_norm": 43.147640228271484,
1037
+ "learning_rate": 7.0003239390994495e-06,
1038
+ "loss": 0.0967,
1039
+ "step": 14200
1040
+ },
1041
+ {
1042
+ "epoch": 4.631578947368421,
1043
+ "grad_norm": 0.5134268999099731,
1044
+ "learning_rate": 6.838354389374798e-06,
1045
+ "loss": 0.0945,
1046
+ "step": 14300
1047
+ },
1048
+ {
1049
+ "epoch": 4.663967611336032,
1050
+ "grad_norm": 0.2741609811782837,
1051
+ "learning_rate": 6.676384839650146e-06,
1052
+ "loss": 0.1012,
1053
+ "step": 14400
1054
+ },
1055
+ {
1056
+ "epoch": 4.696356275303644,
1057
+ "grad_norm": 0.4370046854019165,
1058
+ "learning_rate": 6.514415289925494e-06,
1059
+ "loss": 0.123,
1060
+ "step": 14500
1061
+ },
1062
+ {
1063
+ "epoch": 4.728744939271255,
1064
+ "grad_norm": 4.210660457611084,
1065
+ "learning_rate": 6.352445740200842e-06,
1066
+ "loss": 0.1047,
1067
+ "step": 14600
1068
+ },
1069
+ {
1070
+ "epoch": 4.761133603238866,
1071
+ "grad_norm": 6.13052225112915,
1072
+ "learning_rate": 6.190476190476191e-06,
1073
+ "loss": 0.0966,
1074
+ "step": 14700
1075
+ },
1076
+ {
1077
+ "epoch": 4.793522267206478,
1078
+ "grad_norm": 0.41901007294654846,
1079
+ "learning_rate": 6.028506640751539e-06,
1080
+ "loss": 0.1014,
1081
+ "step": 14800
1082
+ },
1083
+ {
1084
+ "epoch": 4.825910931174089,
1085
+ "grad_norm": 0.28807559609413147,
1086
+ "learning_rate": 5.866537091026887e-06,
1087
+ "loss": 0.1042,
1088
+ "step": 14900
1089
+ },
1090
+ {
1091
+ "epoch": 4.8582995951417,
1092
+ "grad_norm": 28.3045654296875,
1093
+ "learning_rate": 5.704567541302235e-06,
1094
+ "loss": 0.1027,
1095
+ "step": 15000
1096
+ },
1097
+ {
1098
+ "epoch": 4.890688259109312,
1099
+ "grad_norm": 148.7666473388672,
1100
+ "learning_rate": 5.542597991577583e-06,
1101
+ "loss": 0.0804,
1102
+ "step": 15100
1103
+ },
1104
+ {
1105
+ "epoch": 4.923076923076923,
1106
+ "grad_norm": 1.260237216949463,
1107
+ "learning_rate": 5.380628441852932e-06,
1108
+ "loss": 0.1201,
1109
+ "step": 15200
1110
+ },
1111
+ {
1112
+ "epoch": 4.955465587044534,
1113
+ "grad_norm": 0.15356577932834625,
1114
+ "learning_rate": 5.21865889212828e-06,
1115
+ "loss": 0.1002,
1116
+ "step": 15300
1117
+ },
1118
+ {
1119
+ "epoch": 4.987854251012146,
1120
+ "grad_norm": 0.7852919697761536,
1121
+ "learning_rate": 5.056689342403628e-06,
1122
+ "loss": 0.1202,
1123
+ "step": 15400
1124
+ },
1125
+ {
1126
+ "epoch": 4.9998380566801615,
1127
+ "eval_accuracy": 0.9225352112676056,
1128
+ "eval_loss": 0.3642682433128357,
1129
+ "eval_runtime": 4.0522,
1130
+ "eval_samples_per_second": 490.597,
1131
+ "eval_steps_per_second": 61.448,
1132
+ "step": 15437
1133
+ },
1134
+ {
1135
+ "epoch": 5.020242914979757,
1136
+ "grad_norm": 0.10688459873199463,
1137
+ "learning_rate": 4.894719792678976e-06,
1138
+ "loss": 0.1028,
1139
+ "step": 15500
1140
+ },
1141
+ {
1142
+ "epoch": 5.052631578947368,
1143
+ "grad_norm": 20.821151733398438,
1144
+ "learning_rate": 4.7327502429543244e-06,
1145
+ "loss": 0.0861,
1146
+ "step": 15600
1147
+ },
1148
+ {
1149
+ "epoch": 5.08502024291498,
1150
+ "grad_norm": 0.06313851475715637,
1151
+ "learning_rate": 4.5707806932296735e-06,
1152
+ "loss": 0.0749,
1153
+ "step": 15700
1154
+ },
1155
+ {
1156
+ "epoch": 5.117408906882591,
1157
+ "grad_norm": 5.4235734939575195,
1158
+ "learning_rate": 4.408811143505022e-06,
1159
+ "loss": 0.0756,
1160
+ "step": 15800
1161
+ },
1162
+ {
1163
+ "epoch": 5.149797570850202,
1164
+ "grad_norm": 0.07610571384429932,
1165
+ "learning_rate": 4.24684159378037e-06,
1166
+ "loss": 0.0821,
1167
+ "step": 15900
1168
+ },
1169
+ {
1170
+ "epoch": 5.182186234817814,
1171
+ "grad_norm": 189.49761962890625,
1172
+ "learning_rate": 4.084872044055718e-06,
1173
+ "loss": 0.0694,
1174
+ "step": 16000
1175
+ },
1176
+ {
1177
+ "epoch": 5.2145748987854255,
1178
+ "grad_norm": 0.05370008572936058,
1179
+ "learning_rate": 3.922902494331066e-06,
1180
+ "loss": 0.0767,
1181
+ "step": 16100
1182
+ },
1183
+ {
1184
+ "epoch": 5.246963562753036,
1185
+ "grad_norm": 0.05699535831809044,
1186
+ "learning_rate": 3.7609329446064145e-06,
1187
+ "loss": 0.0777,
1188
+ "step": 16200
1189
+ },
1190
+ {
1191
+ "epoch": 5.279352226720648,
1192
+ "grad_norm": 0.04399504140019417,
1193
+ "learning_rate": 3.5989633948817623e-06,
1194
+ "loss": 0.0753,
1195
+ "step": 16300
1196
+ },
1197
+ {
1198
+ "epoch": 5.3117408906882595,
1199
+ "grad_norm": 0.09438109397888184,
1200
+ "learning_rate": 3.4369938451571105e-06,
1201
+ "loss": 0.0759,
1202
+ "step": 16400
1203
+ },
1204
+ {
1205
+ "epoch": 5.34412955465587,
1206
+ "grad_norm": 0.10693158209323883,
1207
+ "learning_rate": 3.2750242954324587e-06,
1208
+ "loss": 0.0627,
1209
+ "step": 16500
1210
+ },
1211
+ {
1212
+ "epoch": 5.376518218623482,
1213
+ "grad_norm": 0.16652892529964447,
1214
+ "learning_rate": 3.113054745707807e-06,
1215
+ "loss": 0.0738,
1216
+ "step": 16600
1217
+ },
1218
+ {
1219
+ "epoch": 5.4089068825910935,
1220
+ "grad_norm": 17.245641708374023,
1221
+ "learning_rate": 2.951085195983155e-06,
1222
+ "loss": 0.0806,
1223
+ "step": 16700
1224
+ },
1225
+ {
1226
+ "epoch": 5.441295546558704,
1227
+ "grad_norm": 0.10547757893800735,
1228
+ "learning_rate": 2.7891156462585034e-06,
1229
+ "loss": 0.0671,
1230
+ "step": 16800
1231
+ },
1232
+ {
1233
+ "epoch": 5.473684210526316,
1234
+ "grad_norm": 1.494895100593567,
1235
+ "learning_rate": 2.6271460965338516e-06,
1236
+ "loss": 0.0733,
1237
+ "step": 16900
1238
+ },
1239
+ {
1240
+ "epoch": 5.506072874493928,
1241
+ "grad_norm": 0.5026708841323853,
1242
+ "learning_rate": 2.4651765468091998e-06,
1243
+ "loss": 0.0579,
1244
+ "step": 17000
1245
+ },
1246
+ {
1247
+ "epoch": 5.538461538461538,
1248
+ "grad_norm": 0.033295173197984695,
1249
+ "learning_rate": 2.303206997084548e-06,
1250
+ "loss": 0.0665,
1251
+ "step": 17100
1252
+ },
1253
+ {
1254
+ "epoch": 5.57085020242915,
1255
+ "grad_norm": 38.09762954711914,
1256
+ "learning_rate": 2.141237447359896e-06,
1257
+ "loss": 0.076,
1258
+ "step": 17200
1259
+ },
1260
+ {
1261
+ "epoch": 5.603238866396762,
1262
+ "grad_norm": 6.565536022186279,
1263
+ "learning_rate": 1.9792678976352444e-06,
1264
+ "loss": 0.0695,
1265
+ "step": 17300
1266
+ },
1267
+ {
1268
+ "epoch": 5.635627530364372,
1269
+ "grad_norm": 0.03646261617541313,
1270
+ "learning_rate": 1.817298347910593e-06,
1271
+ "loss": 0.0941,
1272
+ "step": 17400
1273
+ },
1274
+ {
1275
+ "epoch": 5.668016194331984,
1276
+ "grad_norm": 2.2503933906555176,
1277
+ "learning_rate": 1.6553287981859412e-06,
1278
+ "loss": 0.0992,
1279
+ "step": 17500
1280
+ },
1281
+ {
1282
+ "epoch": 5.700404858299595,
1283
+ "grad_norm": 0.26883581280708313,
1284
+ "learning_rate": 1.4933592484612894e-06,
1285
+ "loss": 0.0849,
1286
+ "step": 17600
1287
+ },
1288
+ {
1289
+ "epoch": 5.732793522267206,
1290
+ "grad_norm": 0.1083364486694336,
1291
+ "learning_rate": 1.3313896987366376e-06,
1292
+ "loss": 0.0798,
1293
+ "step": 17700
1294
+ },
1295
+ {
1296
+ "epoch": 5.765182186234818,
1297
+ "grad_norm": 1.5160281658172607,
1298
+ "learning_rate": 1.1694201490119858e-06,
1299
+ "loss": 0.0535,
1300
+ "step": 17800
1301
+ },
1302
+ {
1303
+ "epoch": 5.797570850202429,
1304
+ "grad_norm": 0.054747115820646286,
1305
+ "learning_rate": 1.007450599287334e-06,
1306
+ "loss": 0.0828,
1307
+ "step": 17900
1308
+ },
1309
+ {
1310
+ "epoch": 5.82995951417004,
1311
+ "grad_norm": 0.062284424901008606,
1312
+ "learning_rate": 8.454810495626823e-07,
1313
+ "loss": 0.0816,
1314
+ "step": 18000
1315
+ },
1316
+ {
1317
+ "epoch": 5.862348178137652,
1318
+ "grad_norm": 0.07957690209150314,
1319
+ "learning_rate": 6.835114998380305e-07,
1320
+ "loss": 0.0873,
1321
+ "step": 18100
1322
+ },
1323
+ {
1324
+ "epoch": 5.894736842105263,
1325
+ "grad_norm": 18.025510787963867,
1326
+ "learning_rate": 5.215419501133787e-07,
1327
+ "loss": 0.0786,
1328
+ "step": 18200
1329
+ },
1330
+ {
1331
+ "epoch": 5.9271255060728745,
1332
+ "grad_norm": 3.286792755126953,
1333
+ "learning_rate": 3.5957240038872693e-07,
1334
+ "loss": 0.0761,
1335
+ "step": 18300
1336
+ },
1337
+ {
1338
+ "epoch": 5.959514170040486,
1339
+ "grad_norm": 0.04595565423369408,
1340
+ "learning_rate": 1.9760285066407517e-07,
1341
+ "loss": 0.0551,
1342
+ "step": 18400
1343
+ },
1344
+ {
1345
+ "epoch": 5.991902834008097,
1346
+ "grad_norm": 0.07384829223155975,
1347
+ "learning_rate": 3.563330093942339e-08,
1348
+ "loss": 0.092,
1349
+ "step": 18500
1350
+ },
1351
+ {
1352
+ "epoch": 5.999028340080971,
1353
+ "eval_accuracy": 0.9164989939637826,
1354
+ "eval_loss": 0.4001348912715912,
1355
+ "eval_runtime": 4.0447,
1356
+ "eval_samples_per_second": 491.512,
1357
+ "eval_steps_per_second": 61.563,
1358
+ "step": 18522
1359
+ },
1360
+ {
1361
+ "epoch": 5.999028340080971,
1362
+ "step": 18522,
1363
+ "total_flos": 1.33712370278538e+16,
1364
+ "train_loss": 0.17084743083006051,
1365
+ "train_runtime": 2225.7449,
1366
+ "train_samples_per_second": 133.166,
1367
+ "train_steps_per_second": 8.322
1368
+ }
1369
+ ],
1370
+ "logging_steps": 100,
1371
+ "max_steps": 18522,
1372
+ "num_input_tokens_seen": 0,
1373
+ "num_train_epochs": 6,
1374
+ "save_steps": 500,
1375
+ "stateful_callbacks": {
1376
+ "TrainerControl": {
1377
+ "args": {
1378
+ "should_epoch_stop": false,
1379
+ "should_evaluate": false,
1380
+ "should_log": false,
1381
+ "should_save": true,
1382
+ "should_training_stop": true
1383
+ },
1384
+ "attributes": {}
1385
+ }
1386
+ },
1387
+ "total_flos": 1.33712370278538e+16,
1388
+ "train_batch_size": 8,
1389
+ "trial_name": null,
1390
+ "trial_params": null
1391
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10f5a410a5dc3ef8d5cd2b65e3990ccb7f39eff14936d78e6e05d2e12dfd7723
3
+ size 5560
vocab.txt ADDED
The diff for this file is too large to render. See raw diff