"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/trainer_state.json +798 -3
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630259719.897293/events.out.tfevents.1630259719.cc93b136ebf5.1086.249 +3 -0
- model-bin/finetune/base/log/1630260157.7758684/events.out.tfevents.1630260157.cc93b136ebf5.1086.251 +3 -0
- model-bin/finetune/base/log/1630260591.4709191/events.out.tfevents.1630260591.cc93b136ebf5.1086.253 +3 -0
- model-bin/finetune/base/log/1630261026.7725165/events.out.tfevents.1630261026.cc93b136ebf5.1086.255 +3 -0
- model-bin/finetune/base/log/1630261460.190783/events.out.tfevents.1630261460.cc93b136ebf5.1086.257 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630259719.cc93b136ebf5.1086.248 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630260157.cc93b136ebf5.1086.250 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630260591.cc93b136ebf5.1086.252 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630261026.cc93b136ebf5.1086.254 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630261460.cc93b136ebf5.1086.256 +3 -0
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/optimizer.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 722165393
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:470c3ea26795136bd282cb4eb7e01b82e7f74de658b159421d45cd8357c5df2e
|
3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/preprocessor_config.json
RENAMED
File without changes
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/pytorch_model.bin
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377909911
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b99febcccd6a1f50c15709024b6aadd57bf558412f09f39446be17d92b116544
|
3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/rng_state.pth
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14503
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:847621cdc9516237f2d0544ce00f3369310442cbe0ff46e33f4438f8246622c7
|
3 |
size 14503
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/scaler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 559
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1adce13f705e5edbbf832291b3d4b862eb38f0d353dd7a919d7e7015308b5f53
|
3 |
size 559
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/scheduler.pt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 623
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ccc55077eebec856cae793ae16f745e98931efbbdb640f5646e5c41e6164ecd5
|
3 |
size 623
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/trainer_state.json
RENAMED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.1668011442822563,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -296010,11 +296010,806 @@
|
|
296010 |
"eval_steps_per_second": 0.654,
|
296011 |
"eval_wer": 0.17398648648648649,
|
296012 |
"step": 154557
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
296013 |
}
|
296014 |
],
|
296015 |
"max_steps": 620000,
|
296016 |
"num_train_epochs": 5000,
|
296017 |
-
"total_flos": 4.
|
296018 |
"trial_name": null,
|
296019 |
"trial_params": null
|
296020 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.1668011442822563,
|
3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
4 |
+
"epoch": 1250.995983935743,
|
5 |
+
"global_step": 155180,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
296010 |
"eval_steps_per_second": 0.654,
|
296011 |
"eval_wer": 0.17398648648648649,
|
296012 |
"step": 154557
|
296013 |
+
},
|
296014 |
+
{
|
296015 |
+
"epoch": 1246.02,
|
296016 |
+
"learning_rate": 7.5202100161550895e-06,
|
296017 |
+
"loss": 0.3359,
|
296018 |
+
"step": 154560
|
296019 |
+
},
|
296020 |
+
{
|
296021 |
+
"epoch": 1246.06,
|
296022 |
+
"learning_rate": 7.520129240710825e-06,
|
296023 |
+
"loss": 0.2438,
|
296024 |
+
"step": 154565
|
296025 |
+
},
|
296026 |
+
{
|
296027 |
+
"epoch": 1246.1,
|
296028 |
+
"learning_rate": 7.5200484652665595e-06,
|
296029 |
+
"loss": 0.2681,
|
296030 |
+
"step": 154570
|
296031 |
+
},
|
296032 |
+
{
|
296033 |
+
"epoch": 1246.14,
|
296034 |
+
"learning_rate": 7.519967689822295e-06,
|
296035 |
+
"loss": 0.3789,
|
296036 |
+
"step": 154575
|
296037 |
+
},
|
296038 |
+
{
|
296039 |
+
"epoch": 1246.18,
|
296040 |
+
"learning_rate": 7.5198869143780295e-06,
|
296041 |
+
"loss": 0.6661,
|
296042 |
+
"step": 154580
|
296043 |
+
},
|
296044 |
+
{
|
296045 |
+
"epoch": 1246.22,
|
296046 |
+
"learning_rate": 7.519806138933765e-06,
|
296047 |
+
"loss": 0.8786,
|
296048 |
+
"step": 154585
|
296049 |
+
},
|
296050 |
+
{
|
296051 |
+
"epoch": 1246.27,
|
296052 |
+
"learning_rate": 7.5197253634894995e-06,
|
296053 |
+
"loss": 0.2525,
|
296054 |
+
"step": 154590
|
296055 |
+
},
|
296056 |
+
{
|
296057 |
+
"epoch": 1246.31,
|
296058 |
+
"learning_rate": 7.519644588045235e-06,
|
296059 |
+
"loss": 0.2866,
|
296060 |
+
"step": 154595
|
296061 |
+
},
|
296062 |
+
{
|
296063 |
+
"epoch": 1246.35,
|
296064 |
+
"learning_rate": 7.5195638126009694e-06,
|
296065 |
+
"loss": 0.3159,
|
296066 |
+
"step": 154600
|
296067 |
+
},
|
296068 |
+
{
|
296069 |
+
"epoch": 1246.39,
|
296070 |
+
"learning_rate": 7.519483037156705e-06,
|
296071 |
+
"loss": 0.6285,
|
296072 |
+
"step": 154605
|
296073 |
+
},
|
296074 |
+
{
|
296075 |
+
"epoch": 1246.43,
|
296076 |
+
"learning_rate": 7.519402261712439e-06,
|
296077 |
+
"loss": 0.8862,
|
296078 |
+
"step": 154610
|
296079 |
+
},
|
296080 |
+
{
|
296081 |
+
"epoch": 1246.47,
|
296082 |
+
"learning_rate": 7.519321486268175e-06,
|
296083 |
+
"loss": 0.2578,
|
296084 |
+
"step": 154615
|
296085 |
+
},
|
296086 |
+
{
|
296087 |
+
"epoch": 1246.51,
|
296088 |
+
"learning_rate": 7.519240710823909e-06,
|
296089 |
+
"loss": 0.2415,
|
296090 |
+
"step": 154620
|
296091 |
+
},
|
296092 |
+
{
|
296093 |
+
"epoch": 1246.55,
|
296094 |
+
"learning_rate": 7.519159935379645e-06,
|
296095 |
+
"loss": 0.3066,
|
296096 |
+
"step": 154625
|
296097 |
+
},
|
296098 |
+
{
|
296099 |
+
"epoch": 1246.59,
|
296100 |
+
"learning_rate": 7.519079159935381e-06,
|
296101 |
+
"loss": 0.5628,
|
296102 |
+
"step": 154630
|
296103 |
+
},
|
296104 |
+
{
|
296105 |
+
"epoch": 1246.63,
|
296106 |
+
"learning_rate": 7.518998384491115e-06,
|
296107 |
+
"loss": 0.865,
|
296108 |
+
"step": 154635
|
296109 |
+
},
|
296110 |
+
{
|
296111 |
+
"epoch": 1246.67,
|
296112 |
+
"learning_rate": 7.518917609046851e-06,
|
296113 |
+
"loss": 0.2865,
|
296114 |
+
"step": 154640
|
296115 |
+
},
|
296116 |
+
{
|
296117 |
+
"epoch": 1246.71,
|
296118 |
+
"learning_rate": 7.518836833602585e-06,
|
296119 |
+
"loss": 0.2618,
|
296120 |
+
"step": 154645
|
296121 |
+
},
|
296122 |
+
{
|
296123 |
+
"epoch": 1246.75,
|
296124 |
+
"learning_rate": 7.518756058158321e-06,
|
296125 |
+
"loss": 0.3727,
|
296126 |
+
"step": 154650
|
296127 |
+
},
|
296128 |
+
{
|
296129 |
+
"epoch": 1246.79,
|
296130 |
+
"learning_rate": 7.518675282714055e-06,
|
296131 |
+
"loss": 0.8115,
|
296132 |
+
"step": 154655
|
296133 |
+
},
|
296134 |
+
{
|
296135 |
+
"epoch": 1246.83,
|
296136 |
+
"learning_rate": 7.518594507269791e-06,
|
296137 |
+
"loss": 0.7482,
|
296138 |
+
"step": 154660
|
296139 |
+
},
|
296140 |
+
{
|
296141 |
+
"epoch": 1246.87,
|
296142 |
+
"learning_rate": 7.518513731825525e-06,
|
296143 |
+
"loss": 0.3001,
|
296144 |
+
"step": 154665
|
296145 |
+
},
|
296146 |
+
{
|
296147 |
+
"epoch": 1246.91,
|
296148 |
+
"learning_rate": 7.518432956381261e-06,
|
296149 |
+
"loss": 0.3185,
|
296150 |
+
"step": 154670
|
296151 |
+
},
|
296152 |
+
{
|
296153 |
+
"epoch": 1246.95,
|
296154 |
+
"learning_rate": 7.518352180936995e-06,
|
296155 |
+
"loss": 0.3938,
|
296156 |
+
"step": 154675
|
296157 |
+
},
|
296158 |
+
{
|
296159 |
+
"epoch": 1246.99,
|
296160 |
+
"learning_rate": 7.518271405492731e-06,
|
296161 |
+
"loss": 0.8753,
|
296162 |
+
"step": 154680
|
296163 |
+
},
|
296164 |
+
{
|
296165 |
+
"epoch": 1247.0,
|
296166 |
+
"eval_loss": 0.3975435793399811,
|
296167 |
+
"eval_runtime": 41.5728,
|
296168 |
+
"eval_samples_per_second": 20.326,
|
296169 |
+
"eval_steps_per_second": 0.649,
|
296170 |
+
"eval_wer": 0.1845796090400407,
|
296171 |
+
"step": 154681
|
296172 |
+
},
|
296173 |
+
{
|
296174 |
+
"epoch": 1237.03,
|
296175 |
+
"learning_rate": 7.518190630048467e-06,
|
296176 |
+
"loss": 0.3444,
|
296177 |
+
"step": 154685
|
296178 |
+
},
|
296179 |
+
{
|
296180 |
+
"epoch": 1237.07,
|
296181 |
+
"learning_rate": 7.518109854604201e-06,
|
296182 |
+
"loss": 0.2737,
|
296183 |
+
"step": 154690
|
296184 |
+
},
|
296185 |
+
{
|
296186 |
+
"epoch": 1237.11,
|
296187 |
+
"learning_rate": 7.518029079159937e-06,
|
296188 |
+
"loss": 0.2676,
|
296189 |
+
"step": 154695
|
296190 |
+
},
|
296191 |
+
{
|
296192 |
+
"epoch": 1237.15,
|
296193 |
+
"learning_rate": 7.517948303715671e-06,
|
296194 |
+
"loss": 0.3919,
|
296195 |
+
"step": 154700
|
296196 |
+
},
|
296197 |
+
{
|
296198 |
+
"epoch": 1237.19,
|
296199 |
+
"learning_rate": 7.517867528271407e-06,
|
296200 |
+
"loss": 1.0205,
|
296201 |
+
"step": 154705
|
296202 |
+
},
|
296203 |
+
{
|
296204 |
+
"epoch": 1237.23,
|
296205 |
+
"learning_rate": 7.517786752827141e-06,
|
296206 |
+
"loss": 0.6594,
|
296207 |
+
"step": 154710
|
296208 |
+
},
|
296209 |
+
{
|
296210 |
+
"epoch": 1237.27,
|
296211 |
+
"learning_rate": 7.517705977382877e-06,
|
296212 |
+
"loss": 0.2955,
|
296213 |
+
"step": 154715
|
296214 |
+
},
|
296215 |
+
{
|
296216 |
+
"epoch": 1237.31,
|
296217 |
+
"learning_rate": 7.517625201938611e-06,
|
296218 |
+
"loss": 0.2682,
|
296219 |
+
"step": 154720
|
296220 |
+
},
|
296221 |
+
{
|
296222 |
+
"epoch": 1237.35,
|
296223 |
+
"learning_rate": 7.517544426494347e-06,
|
296224 |
+
"loss": 0.3546,
|
296225 |
+
"step": 154725
|
296226 |
+
},
|
296227 |
+
{
|
296228 |
+
"epoch": 1237.39,
|
296229 |
+
"learning_rate": 7.517463651050081e-06,
|
296230 |
+
"loss": 0.8458,
|
296231 |
+
"step": 154730
|
296232 |
+
},
|
296233 |
+
{
|
296234 |
+
"epoch": 1237.43,
|
296235 |
+
"learning_rate": 7.517382875605817e-06,
|
296236 |
+
"loss": 0.6329,
|
296237 |
+
"step": 154735
|
296238 |
+
},
|
296239 |
+
{
|
296240 |
+
"epoch": 1237.47,
|
296241 |
+
"learning_rate": 7.517302100161552e-06,
|
296242 |
+
"loss": 0.2381,
|
296243 |
+
"step": 154740
|
296244 |
+
},
|
296245 |
+
{
|
296246 |
+
"epoch": 1237.51,
|
296247 |
+
"learning_rate": 7.517221324717287e-06,
|
296248 |
+
"loss": 0.3116,
|
296249 |
+
"step": 154745
|
296250 |
+
},
|
296251 |
+
{
|
296252 |
+
"epoch": 1237.55,
|
296253 |
+
"learning_rate": 7.5171405492730225e-06,
|
296254 |
+
"loss": 0.3383,
|
296255 |
+
"step": 154750
|
296256 |
+
},
|
296257 |
+
{
|
296258 |
+
"epoch": 1237.59,
|
296259 |
+
"learning_rate": 7.517059773828757e-06,
|
296260 |
+
"loss": 0.8708,
|
296261 |
+
"step": 154755
|
296262 |
+
},
|
296263 |
+
{
|
296264 |
+
"epoch": 1237.63,
|
296265 |
+
"learning_rate": 7.5169789983844924e-06,
|
296266 |
+
"loss": 0.5733,
|
296267 |
+
"step": 154760
|
296268 |
+
},
|
296269 |
+
{
|
296270 |
+
"epoch": 1237.67,
|
296271 |
+
"learning_rate": 7.5168982229402266e-06,
|
296272 |
+
"loss": 0.247,
|
296273 |
+
"step": 154765
|
296274 |
+
},
|
296275 |
+
{
|
296276 |
+
"epoch": 1237.71,
|
296277 |
+
"learning_rate": 7.516817447495962e-06,
|
296278 |
+
"loss": 0.3016,
|
296279 |
+
"step": 154770
|
296280 |
+
},
|
296281 |
+
{
|
296282 |
+
"epoch": 1237.75,
|
296283 |
+
"learning_rate": 7.5167366720516965e-06,
|
296284 |
+
"loss": 0.4686,
|
296285 |
+
"step": 154775
|
296286 |
+
},
|
296287 |
+
{
|
296288 |
+
"epoch": 1237.79,
|
296289 |
+
"learning_rate": 7.516655896607432e-06,
|
296290 |
+
"loss": 0.8122,
|
296291 |
+
"step": 154780
|
296292 |
+
},
|
296293 |
+
{
|
296294 |
+
"epoch": 1237.83,
|
296295 |
+
"learning_rate": 7.5165751211631665e-06,
|
296296 |
+
"loss": 0.5604,
|
296297 |
+
"step": 154785
|
296298 |
+
},
|
296299 |
+
{
|
296300 |
+
"epoch": 1237.87,
|
296301 |
+
"learning_rate": 7.516494345718902e-06,
|
296302 |
+
"loss": 0.3368,
|
296303 |
+
"step": 154790
|
296304 |
+
},
|
296305 |
+
{
|
296306 |
+
"epoch": 1237.91,
|
296307 |
+
"learning_rate": 7.5164135702746365e-06,
|
296308 |
+
"loss": 0.2458,
|
296309 |
+
"step": 154795
|
296310 |
+
},
|
296311 |
+
{
|
296312 |
+
"epoch": 1237.95,
|
296313 |
+
"learning_rate": 7.516332794830372e-06,
|
296314 |
+
"loss": 0.3618,
|
296315 |
+
"step": 154800
|
296316 |
+
},
|
296317 |
+
{
|
296318 |
+
"epoch": 1237.99,
|
296319 |
+
"learning_rate": 7.516252019386107e-06,
|
296320 |
+
"loss": 0.931,
|
296321 |
+
"step": 154805
|
296322 |
+
},
|
296323 |
+
{
|
296324 |
+
"epoch": 1238.0,
|
296325 |
+
"eval_loss": 0.42138466238975525,
|
296326 |
+
"eval_runtime": 40.226,
|
296327 |
+
"eval_samples_per_second": 20.981,
|
296328 |
+
"eval_steps_per_second": 0.671,
|
296329 |
+
"eval_wer": 0.17503371796793046,
|
296330 |
+
"step": 154806
|
296331 |
+
},
|
296332 |
+
{
|
296333 |
+
"epoch": 1238.03,
|
296334 |
+
"learning_rate": 7.516171243941842e-06,
|
296335 |
+
"loss": 0.2782,
|
296336 |
+
"step": 154810
|
296337 |
+
},
|
296338 |
+
{
|
296339 |
+
"epoch": 1238.07,
|
296340 |
+
"learning_rate": 7.516090468497578e-06,
|
296341 |
+
"loss": 0.2432,
|
296342 |
+
"step": 154815
|
296343 |
+
},
|
296344 |
+
{
|
296345 |
+
"epoch": 1238.11,
|
296346 |
+
"learning_rate": 7.516009693053312e-06,
|
296347 |
+
"loss": 0.2524,
|
296348 |
+
"step": 154820
|
296349 |
+
},
|
296350 |
+
{
|
296351 |
+
"epoch": 1238.15,
|
296352 |
+
"learning_rate": 7.515928917609048e-06,
|
296353 |
+
"loss": 0.3604,
|
296354 |
+
"step": 154825
|
296355 |
+
},
|
296356 |
+
{
|
296357 |
+
"epoch": 1238.19,
|
296358 |
+
"learning_rate": 7.515848142164782e-06,
|
296359 |
+
"loss": 0.934,
|
296360 |
+
"step": 154830
|
296361 |
+
},
|
296362 |
+
{
|
296363 |
+
"epoch": 1238.23,
|
296364 |
+
"learning_rate": 7.515767366720518e-06,
|
296365 |
+
"loss": 0.5878,
|
296366 |
+
"step": 154835
|
296367 |
+
},
|
296368 |
+
{
|
296369 |
+
"epoch": 1238.27,
|
296370 |
+
"learning_rate": 7.515686591276252e-06,
|
296371 |
+
"loss": 0.2449,
|
296372 |
+
"step": 154840
|
296373 |
+
},
|
296374 |
+
{
|
296375 |
+
"epoch": 1238.31,
|
296376 |
+
"learning_rate": 7.515605815831988e-06,
|
296377 |
+
"loss": 0.2886,
|
296378 |
+
"step": 154845
|
296379 |
+
},
|
296380 |
+
{
|
296381 |
+
"epoch": 1238.35,
|
296382 |
+
"learning_rate": 7.515525040387722e-06,
|
296383 |
+
"loss": 0.3381,
|
296384 |
+
"step": 154850
|
296385 |
+
},
|
296386 |
+
{
|
296387 |
+
"epoch": 1238.39,
|
296388 |
+
"learning_rate": 7.515444264943458e-06,
|
296389 |
+
"loss": 0.7834,
|
296390 |
+
"step": 154855
|
296391 |
+
},
|
296392 |
+
{
|
296393 |
+
"epoch": 1238.43,
|
296394 |
+
"learning_rate": 7.515363489499193e-06,
|
296395 |
+
"loss": 0.7959,
|
296396 |
+
"step": 154860
|
296397 |
+
},
|
296398 |
+
{
|
296399 |
+
"epoch": 1238.47,
|
296400 |
+
"learning_rate": 7.515282714054928e-06,
|
296401 |
+
"loss": 0.2812,
|
296402 |
+
"step": 154865
|
296403 |
+
},
|
296404 |
+
{
|
296405 |
+
"epoch": 1238.51,
|
296406 |
+
"learning_rate": 7.515201938610663e-06,
|
296407 |
+
"loss": 0.2775,
|
296408 |
+
"step": 154870
|
296409 |
+
},
|
296410 |
+
{
|
296411 |
+
"epoch": 1238.55,
|
296412 |
+
"learning_rate": 7.515121163166398e-06,
|
296413 |
+
"loss": 0.4103,
|
296414 |
+
"step": 154875
|
296415 |
+
},
|
296416 |
+
{
|
296417 |
+
"epoch": 1238.59,
|
296418 |
+
"learning_rate": 7.515040387722134e-06,
|
296419 |
+
"loss": 0.8922,
|
296420 |
+
"step": 154880
|
296421 |
+
},
|
296422 |
+
{
|
296423 |
+
"epoch": 1238.63,
|
296424 |
+
"learning_rate": 7.514959612277868e-06,
|
296425 |
+
"loss": 0.696,
|
296426 |
+
"step": 154885
|
296427 |
+
},
|
296428 |
+
{
|
296429 |
+
"epoch": 1238.67,
|
296430 |
+
"learning_rate": 7.514878836833604e-06,
|
296431 |
+
"loss": 0.2623,
|
296432 |
+
"step": 154890
|
296433 |
+
},
|
296434 |
+
{
|
296435 |
+
"epoch": 1238.71,
|
296436 |
+
"learning_rate": 7.514798061389338e-06,
|
296437 |
+
"loss": 0.2932,
|
296438 |
+
"step": 154895
|
296439 |
+
},
|
296440 |
+
{
|
296441 |
+
"epoch": 1238.75,
|
296442 |
+
"learning_rate": 7.514717285945074e-06,
|
296443 |
+
"loss": 0.3314,
|
296444 |
+
"step": 154900
|
296445 |
+
},
|
296446 |
+
{
|
296447 |
+
"epoch": 1238.79,
|
296448 |
+
"learning_rate": 7.514636510500808e-06,
|
296449 |
+
"loss": 0.8592,
|
296450 |
+
"step": 154905
|
296451 |
+
},
|
296452 |
+
{
|
296453 |
+
"epoch": 1238.83,
|
296454 |
+
"learning_rate": 7.514555735056544e-06,
|
296455 |
+
"loss": 0.5791,
|
296456 |
+
"step": 154910
|
296457 |
+
},
|
296458 |
+
{
|
296459 |
+
"epoch": 1238.87,
|
296460 |
+
"learning_rate": 7.514474959612279e-06,
|
296461 |
+
"loss": 0.2748,
|
296462 |
+
"step": 154915
|
296463 |
+
},
|
296464 |
+
{
|
296465 |
+
"epoch": 1238.91,
|
296466 |
+
"learning_rate": 7.514394184168014e-06,
|
296467 |
+
"loss": 0.3728,
|
296468 |
+
"step": 154920
|
296469 |
+
},
|
296470 |
+
{
|
296471 |
+
"epoch": 1238.95,
|
296472 |
+
"learning_rate": 7.514313408723749e-06,
|
296473 |
+
"loss": 0.4352,
|
296474 |
+
"step": 154925
|
296475 |
+
},
|
296476 |
+
{
|
296477 |
+
"epoch": 1238.99,
|
296478 |
+
"learning_rate": 7.514232633279484e-06,
|
296479 |
+
"loss": 0.8688,
|
296480 |
+
"step": 154930
|
296481 |
+
},
|
296482 |
+
{
|
296483 |
+
"epoch": 1239.0,
|
296484 |
+
"eval_loss": 0.36569637060165405,
|
296485 |
+
"eval_runtime": 40.9116,
|
296486 |
+
"eval_samples_per_second": 20.654,
|
296487 |
+
"eval_steps_per_second": 0.66,
|
296488 |
+
"eval_wer": 0.18007352941176472,
|
296489 |
+
"step": 154931
|
296490 |
+
},
|
296491 |
+
{
|
296492 |
+
"epoch": 1239.03,
|
296493 |
+
"learning_rate": 7.514151857835219e-06,
|
296494 |
+
"loss": 0.3529,
|
296495 |
+
"step": 154935
|
296496 |
+
},
|
296497 |
+
{
|
296498 |
+
"epoch": 1239.07,
|
296499 |
+
"learning_rate": 7.514071082390954e-06,
|
296500 |
+
"loss": 0.2858,
|
296501 |
+
"step": 154940
|
296502 |
+
},
|
296503 |
+
{
|
296504 |
+
"epoch": 1239.11,
|
296505 |
+
"learning_rate": 7.513990306946689e-06,
|
296506 |
+
"loss": 0.3375,
|
296507 |
+
"step": 154945
|
296508 |
+
},
|
296509 |
+
{
|
296510 |
+
"epoch": 1239.15,
|
296511 |
+
"learning_rate": 7.513909531502424e-06,
|
296512 |
+
"loss": 0.3088,
|
296513 |
+
"step": 154950
|
296514 |
+
},
|
296515 |
+
{
|
296516 |
+
"epoch": 1239.19,
|
296517 |
+
"learning_rate": 7.5138287560581595e-06,
|
296518 |
+
"loss": 0.7962,
|
296519 |
+
"step": 154955
|
296520 |
+
},
|
296521 |
+
{
|
296522 |
+
"epoch": 1239.23,
|
296523 |
+
"learning_rate": 7.513747980613894e-06,
|
296524 |
+
"loss": 0.6496,
|
296525 |
+
"step": 154960
|
296526 |
+
},
|
296527 |
+
{
|
296528 |
+
"epoch": 1239.27,
|
296529 |
+
"learning_rate": 7.5136672051696295e-06,
|
296530 |
+
"loss": 0.3379,
|
296531 |
+
"step": 154965
|
296532 |
+
},
|
296533 |
+
{
|
296534 |
+
"epoch": 1239.31,
|
296535 |
+
"learning_rate": 7.513586429725364e-06,
|
296536 |
+
"loss": 0.2975,
|
296537 |
+
"step": 154970
|
296538 |
+
},
|
296539 |
+
{
|
296540 |
+
"epoch": 1239.35,
|
296541 |
+
"learning_rate": 7.5135056542810994e-06,
|
296542 |
+
"loss": 0.3822,
|
296543 |
+
"step": 154975
|
296544 |
+
},
|
296545 |
+
{
|
296546 |
+
"epoch": 1239.39,
|
296547 |
+
"learning_rate": 7.513424878836834e-06,
|
296548 |
+
"loss": 0.9315,
|
296549 |
+
"step": 154980
|
296550 |
+
},
|
296551 |
+
{
|
296552 |
+
"epoch": 1239.43,
|
296553 |
+
"learning_rate": 7.513344103392569e-06,
|
296554 |
+
"loss": 0.6303,
|
296555 |
+
"step": 154985
|
296556 |
+
},
|
296557 |
+
{
|
296558 |
+
"epoch": 1239.47,
|
296559 |
+
"learning_rate": 7.513263327948304e-06,
|
296560 |
+
"loss": 0.3489,
|
296561 |
+
"step": 154990
|
296562 |
+
},
|
296563 |
+
{
|
296564 |
+
"epoch": 1239.51,
|
296565 |
+
"learning_rate": 7.513182552504039e-06,
|
296566 |
+
"loss": 0.3152,
|
296567 |
+
"step": 154995
|
296568 |
+
},
|
296569 |
+
{
|
296570 |
+
"epoch": 1239.55,
|
296571 |
+
"learning_rate": 7.513101777059774e-06,
|
296572 |
+
"loss": 0.3954,
|
296573 |
+
"step": 155000
|
296574 |
+
},
|
296575 |
+
{
|
296576 |
+
"epoch": 1239.59,
|
296577 |
+
"learning_rate": 7.513021001615509e-06,
|
296578 |
+
"loss": 0.878,
|
296579 |
+
"step": 155005
|
296580 |
+
},
|
296581 |
+
{
|
296582 |
+
"epoch": 1239.63,
|
296583 |
+
"learning_rate": 7.512940226171244e-06,
|
296584 |
+
"loss": 0.627,
|
296585 |
+
"step": 155010
|
296586 |
+
},
|
296587 |
+
{
|
296588 |
+
"epoch": 1239.67,
|
296589 |
+
"learning_rate": 7.512859450726979e-06,
|
296590 |
+
"loss": 0.2331,
|
296591 |
+
"step": 155015
|
296592 |
+
},
|
296593 |
+
{
|
296594 |
+
"epoch": 1239.71,
|
296595 |
+
"learning_rate": 7.512778675282715e-06,
|
296596 |
+
"loss": 0.2644,
|
296597 |
+
"step": 155020
|
296598 |
+
},
|
296599 |
+
{
|
296600 |
+
"epoch": 1239.75,
|
296601 |
+
"learning_rate": 7.512697899838449e-06,
|
296602 |
+
"loss": 0.3715,
|
296603 |
+
"step": 155025
|
296604 |
+
},
|
296605 |
+
{
|
296606 |
+
"epoch": 1239.79,
|
296607 |
+
"learning_rate": 7.512617124394185e-06,
|
296608 |
+
"loss": 0.9216,
|
296609 |
+
"step": 155030
|
296610 |
+
},
|
296611 |
+
{
|
296612 |
+
"epoch": 1239.83,
|
296613 |
+
"learning_rate": 7.51253634894992e-06,
|
296614 |
+
"loss": 0.6866,
|
296615 |
+
"step": 155035
|
296616 |
+
},
|
296617 |
+
{
|
296618 |
+
"epoch": 1239.87,
|
296619 |
+
"learning_rate": 7.512455573505655e-06,
|
296620 |
+
"loss": 0.2765,
|
296621 |
+
"step": 155040
|
296622 |
+
},
|
296623 |
+
{
|
296624 |
+
"epoch": 1239.91,
|
296625 |
+
"learning_rate": 7.51237479806139e-06,
|
296626 |
+
"loss": 0.2844,
|
296627 |
+
"step": 155045
|
296628 |
+
},
|
296629 |
+
{
|
296630 |
+
"epoch": 1239.95,
|
296631 |
+
"learning_rate": 7.512294022617125e-06,
|
296632 |
+
"loss": 0.3856,
|
296633 |
+
"step": 155050
|
296634 |
+
},
|
296635 |
+
{
|
296636 |
+
"epoch": 1239.99,
|
296637 |
+
"learning_rate": 7.51221324717286e-06,
|
296638 |
+
"loss": 1.0202,
|
296639 |
+
"step": 155055
|
296640 |
+
},
|
296641 |
+
{
|
296642 |
+
"epoch": 1240.0,
|
296643 |
+
"eval_loss": 0.30646491050720215,
|
296644 |
+
"eval_runtime": 40.3619,
|
296645 |
+
"eval_samples_per_second": 20.936,
|
296646 |
+
"eval_steps_per_second": 0.669,
|
296647 |
+
"eval_wer": 0.17656983568075119,
|
296648 |
+
"step": 155056
|
296649 |
+
},
|
296650 |
+
{
|
296651 |
+
"epoch": 1250.03,
|
296652 |
+
"learning_rate": 7.512132471728595e-06,
|
296653 |
+
"loss": 0.2716,
|
296654 |
+
"step": 155060
|
296655 |
+
},
|
296656 |
+
{
|
296657 |
+
"epoch": 1250.07,
|
296658 |
+
"learning_rate": 7.51205169628433e-06,
|
296659 |
+
"loss": 0.2815,
|
296660 |
+
"step": 155065
|
296661 |
+
},
|
296662 |
+
{
|
296663 |
+
"epoch": 1250.11,
|
296664 |
+
"learning_rate": 7.511970920840065e-06,
|
296665 |
+
"loss": 0.3168,
|
296666 |
+
"step": 155070
|
296667 |
+
},
|
296668 |
+
{
|
296669 |
+
"epoch": 1250.15,
|
296670 |
+
"learning_rate": 7.5118901453958e-06,
|
296671 |
+
"loss": 0.3727,
|
296672 |
+
"step": 155075
|
296673 |
+
},
|
296674 |
+
{
|
296675 |
+
"epoch": 1250.19,
|
296676 |
+
"learning_rate": 7.511809369951535e-06,
|
296677 |
+
"loss": 0.9729,
|
296678 |
+
"step": 155080
|
296679 |
+
},
|
296680 |
+
{
|
296681 |
+
"epoch": 1250.23,
|
296682 |
+
"learning_rate": 7.511728594507271e-06,
|
296683 |
+
"loss": 0.7132,
|
296684 |
+
"step": 155085
|
296685 |
+
},
|
296686 |
+
{
|
296687 |
+
"epoch": 1250.27,
|
296688 |
+
"learning_rate": 7.511647819063006e-06,
|
296689 |
+
"loss": 0.3155,
|
296690 |
+
"step": 155090
|
296691 |
+
},
|
296692 |
+
{
|
296693 |
+
"epoch": 1250.31,
|
296694 |
+
"learning_rate": 7.511567043618741e-06,
|
296695 |
+
"loss": 0.3005,
|
296696 |
+
"step": 155095
|
296697 |
+
},
|
296698 |
+
{
|
296699 |
+
"epoch": 1250.35,
|
296700 |
+
"learning_rate": 7.511486268174476e-06,
|
296701 |
+
"loss": 0.3456,
|
296702 |
+
"step": 155100
|
296703 |
+
},
|
296704 |
+
{
|
296705 |
+
"epoch": 1250.39,
|
296706 |
+
"learning_rate": 7.511405492730211e-06,
|
296707 |
+
"loss": 0.7762,
|
296708 |
+
"step": 155105
|
296709 |
+
},
|
296710 |
+
{
|
296711 |
+
"epoch": 1250.43,
|
296712 |
+
"learning_rate": 7.511324717285946e-06,
|
296713 |
+
"loss": 0.6201,
|
296714 |
+
"step": 155110
|
296715 |
+
},
|
296716 |
+
{
|
296717 |
+
"epoch": 1250.47,
|
296718 |
+
"learning_rate": 7.511243941841681e-06,
|
296719 |
+
"loss": 0.2818,
|
296720 |
+
"step": 155115
|
296721 |
+
},
|
296722 |
+
{
|
296723 |
+
"epoch": 1250.51,
|
296724 |
+
"learning_rate": 7.511163166397416e-06,
|
296725 |
+
"loss": 0.2924,
|
296726 |
+
"step": 155120
|
296727 |
+
},
|
296728 |
+
{
|
296729 |
+
"epoch": 1250.55,
|
296730 |
+
"learning_rate": 7.511082390953151e-06,
|
296731 |
+
"loss": 0.3261,
|
296732 |
+
"step": 155125
|
296733 |
+
},
|
296734 |
+
{
|
296735 |
+
"epoch": 1250.59,
|
296736 |
+
"learning_rate": 7.511001615508886e-06,
|
296737 |
+
"loss": 0.8687,
|
296738 |
+
"step": 155130
|
296739 |
+
},
|
296740 |
+
{
|
296741 |
+
"epoch": 1250.63,
|
296742 |
+
"learning_rate": 7.510920840064621e-06,
|
296743 |
+
"loss": 0.5005,
|
296744 |
+
"step": 155135
|
296745 |
+
},
|
296746 |
+
{
|
296747 |
+
"epoch": 1250.67,
|
296748 |
+
"learning_rate": 7.510840064620356e-06,
|
296749 |
+
"loss": 0.2698,
|
296750 |
+
"step": 155140
|
296751 |
+
},
|
296752 |
+
{
|
296753 |
+
"epoch": 1250.71,
|
296754 |
+
"learning_rate": 7.510759289176091e-06,
|
296755 |
+
"loss": 0.2844,
|
296756 |
+
"step": 155145
|
296757 |
+
},
|
296758 |
+
{
|
296759 |
+
"epoch": 1250.76,
|
296760 |
+
"learning_rate": 7.510678513731826e-06,
|
296761 |
+
"loss": 0.3468,
|
296762 |
+
"step": 155150
|
296763 |
+
},
|
296764 |
+
{
|
296765 |
+
"epoch": 1250.8,
|
296766 |
+
"learning_rate": 7.5105977382875615e-06,
|
296767 |
+
"loss": 0.9883,
|
296768 |
+
"step": 155155
|
296769 |
+
},
|
296770 |
+
{
|
296771 |
+
"epoch": 1250.84,
|
296772 |
+
"learning_rate": 7.5105169628432965e-06,
|
296773 |
+
"loss": 0.6705,
|
296774 |
+
"step": 155160
|
296775 |
+
},
|
296776 |
+
{
|
296777 |
+
"epoch": 1250.88,
|
296778 |
+
"learning_rate": 7.5104361873990315e-06,
|
296779 |
+
"loss": 0.2834,
|
296780 |
+
"step": 155165
|
296781 |
+
},
|
296782 |
+
{
|
296783 |
+
"epoch": 1250.92,
|
296784 |
+
"learning_rate": 7.5103554119547665e-06,
|
296785 |
+
"loss": 0.3121,
|
296786 |
+
"step": 155170
|
296787 |
+
},
|
296788 |
+
{
|
296789 |
+
"epoch": 1250.96,
|
296790 |
+
"learning_rate": 7.5102746365105015e-06,
|
296791 |
+
"loss": 0.4157,
|
296792 |
+
"step": 155175
|
296793 |
+
},
|
296794 |
+
{
|
296795 |
+
"epoch": 1251.0,
|
296796 |
+
"learning_rate": 7.5101938610662365e-06,
|
296797 |
+
"loss": 0.9646,
|
296798 |
+
"step": 155180
|
296799 |
+
},
|
296800 |
+
{
|
296801 |
+
"epoch": 1251.0,
|
296802 |
+
"eval_loss": 0.518424391746521,
|
296803 |
+
"eval_runtime": 41.1845,
|
296804 |
+
"eval_samples_per_second": 20.517,
|
296805 |
+
"eval_steps_per_second": 0.656,
|
296806 |
+
"eval_wer": 0.18026912388474478,
|
296807 |
+
"step": 155180
|
296808 |
}
|
296809 |
],
|
296810 |
"max_steps": 620000,
|
296811 |
"num_train_epochs": 5000,
|
296812 |
+
"total_flos": 4.366899653515725e+20,
|
296813 |
"trial_name": null,
|
296814 |
"trial_params": null
|
296815 |
}
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/training_args.bin
RENAMED
File without changes
|
model-bin/finetune/base/log/1630259719.897293/events.out.tfevents.1630259719.cc93b136ebf5.1086.249
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b026cb9d789aece99a3e446b51eef6d6eafe7463ff3affb3acdff973c9cecafb
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630260157.7758684/events.out.tfevents.1630260157.cc93b136ebf5.1086.251
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eb971f3865430bebfa1d58eda62f42d75ec70030bc30ab5b02fcf85757e2658f
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630260591.4709191/events.out.tfevents.1630260591.cc93b136ebf5.1086.253
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6a1abfdef8f5e60f550201561885525bc0f08b046f43b14dd16bd319ab43f679
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630261026.7725165/events.out.tfevents.1630261026.cc93b136ebf5.1086.255
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1e4d315fde2bca0802d4b45f0f764ad61cbd51cadef039b9a244dc44c1e01520
|
3 |
+
size 4194
|
model-bin/finetune/base/log/1630261460.190783/events.out.tfevents.1630261460.cc93b136ebf5.1086.257
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:620ac2c75253b61fa00bf008d66e61a212b0eeb51edec85242ba1b84ca784dd7
|
3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630259719.cc93b136ebf5.1086.248
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f8e93d8cc698e64dca2ac9e57763f5c78feb18455eabfac9c1e2580edc192eb2
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630260157.cc93b136ebf5.1086.250
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fb210509e463781ed894247a0beaf3c1bbd3f02168ab0b5313909da570cfed24
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630260591.cc93b136ebf5.1086.252
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1cac3e5992df5b67b83acf2fc649f479413bf5f72958cbdefca5ba2ed0761bc1
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630261026.cc93b136ebf5.1086.254
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:42bd2ba8b85647d66baebed9e3e1cff725f407db4785030a2cf15ecbc6bdb6ca
|
3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630261460.cc93b136ebf5.1086.256
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:541f0e49276ec7aa96edb6c944e0410f0b0b1cd347b64a03ba4fdbd5c0b0d642
|
3 |
+
size 8622
|