|
{ |
|
"best_metric": 0.45234861969947815, |
|
"best_model_checkpoint": "qlora-extractor-checkpoints/checkpoint-598", |
|
"epoch": 2.392, |
|
"eval_steps": 299, |
|
"global_step": 598, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.10648680478334427, |
|
"learning_rate": 0.00019786096256684492, |
|
"loss": 1.2225, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.0971209928393364, |
|
"learning_rate": 0.00019518716577540107, |
|
"loss": 0.9559, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.120168536901474, |
|
"learning_rate": 0.00019251336898395722, |
|
"loss": 0.7891, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.13561870157718658, |
|
"learning_rate": 0.0001898395721925134, |
|
"loss": 0.6716, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.13140155375003815, |
|
"learning_rate": 0.0001871657754010695, |
|
"loss": 0.5196, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.12888653576374054, |
|
"learning_rate": 0.0001844919786096257, |
|
"loss": 0.8715, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.17288090288639069, |
|
"learning_rate": 0.00018181818181818183, |
|
"loss": 0.7102, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.19013544917106628, |
|
"learning_rate": 0.00017914438502673798, |
|
"loss": 0.6303, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.16771665215492249, |
|
"learning_rate": 0.00017647058823529413, |
|
"loss": 0.5402, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.12465972453355789, |
|
"learning_rate": 0.00017379679144385028, |
|
"loss": 0.4714, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.1407775729894638, |
|
"learning_rate": 0.00017112299465240642, |
|
"loss": 0.7708, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.1658485233783722, |
|
"learning_rate": 0.00016844919786096257, |
|
"loss": 0.5489, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.18059569597244263, |
|
"learning_rate": 0.00016577540106951872, |
|
"loss": 0.6152, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.16564571857452393, |
|
"learning_rate": 0.0001631016042780749, |
|
"loss": 0.5372, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.15012599527835846, |
|
"learning_rate": 0.00016042780748663101, |
|
"loss": 0.4824, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.18662799894809723, |
|
"learning_rate": 0.0001577540106951872, |
|
"loss": 0.7024, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.26714515686035156, |
|
"learning_rate": 0.0001550802139037433, |
|
"loss": 0.6077, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.14682364463806152, |
|
"learning_rate": 0.00015240641711229948, |
|
"loss": 0.5559, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.20032133162021637, |
|
"learning_rate": 0.00014973262032085563, |
|
"loss": 0.5333, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.16041621565818787, |
|
"learning_rate": 0.00014705882352941178, |
|
"loss": 0.4873, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.15586546063423157, |
|
"learning_rate": 0.00014438502673796793, |
|
"loss": 0.6419, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.1745934784412384, |
|
"learning_rate": 0.00014171122994652405, |
|
"loss": 0.5824, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.15679331123828888, |
|
"learning_rate": 0.00013903743315508022, |
|
"loss": 0.5084, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.16343417763710022, |
|
"learning_rate": 0.00013636363636363637, |
|
"loss": 0.481, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.14305788278579712, |
|
"learning_rate": 0.00013368983957219252, |
|
"loss": 0.4472, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.20606954395771027, |
|
"learning_rate": 0.00013101604278074866, |
|
"loss": 0.5238, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.1933826059103012, |
|
"learning_rate": 0.0001283422459893048, |
|
"loss": 0.4699, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 0.1788238286972046, |
|
"learning_rate": 0.00012566844919786096, |
|
"loss": 0.4909, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 0.2144651561975479, |
|
"learning_rate": 0.00012299465240641713, |
|
"loss": 0.387, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 0.18218065798282623, |
|
"learning_rate": 0.00012005347593582887, |
|
"loss": 0.319, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.14094972610473633, |
|
"learning_rate": 0.00011737967914438503, |
|
"loss": 0.6182, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.20308974385261536, |
|
"learning_rate": 0.00011470588235294118, |
|
"loss": 0.4492, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 0.18815161287784576, |
|
"learning_rate": 0.00011203208556149734, |
|
"loss": 0.4472, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.3599999999999999, |
|
"grad_norm": 0.1986108422279358, |
|
"learning_rate": 0.00010935828877005347, |
|
"loss": 0.3996, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 0.1628069281578064, |
|
"learning_rate": 0.00010668449197860964, |
|
"loss": 0.3645, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.14933669567108154, |
|
"learning_rate": 0.00010401069518716577, |
|
"loss": 0.5977, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.18734735250473022, |
|
"learning_rate": 0.00010133689839572193, |
|
"loss": 0.3784, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 0.2121153324842453, |
|
"learning_rate": 9.866310160427808e-05, |
|
"loss": 0.4581, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 0.19748178124427795, |
|
"learning_rate": 9.598930481283423e-05, |
|
"loss": 0.3858, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.18506589531898499, |
|
"learning_rate": 9.331550802139037e-05, |
|
"loss": 0.3706, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.6400000000000001, |
|
"grad_norm": 0.15329745411872864, |
|
"learning_rate": 9.064171122994652e-05, |
|
"loss": 0.5347, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.6800000000000002, |
|
"grad_norm": 0.19260501861572266, |
|
"learning_rate": 8.796791443850267e-05, |
|
"loss": 0.4812, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.1763940006494522, |
|
"learning_rate": 8.529411764705883e-05, |
|
"loss": 0.4317, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 0.20417028665542603, |
|
"learning_rate": 8.262032085561498e-05, |
|
"loss": 0.4145, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 0.2001723051071167, |
|
"learning_rate": 7.994652406417112e-05, |
|
"loss": 0.3884, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.8399999999999999, |
|
"grad_norm": 0.2101861983537674, |
|
"learning_rate": 7.727272727272727e-05, |
|
"loss": 0.4758, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.2131994664669037, |
|
"learning_rate": 7.459893048128342e-05, |
|
"loss": 0.466, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 0.21261604130268097, |
|
"learning_rate": 7.192513368983958e-05, |
|
"loss": 0.4088, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.21039697527885437, |
|
"learning_rate": 6.925133689839573e-05, |
|
"loss": 0.3858, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.18674618005752563, |
|
"learning_rate": 6.657754010695188e-05, |
|
"loss": 0.3672, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"grad_norm": 0.1681678146123886, |
|
"learning_rate": 6.390374331550802e-05, |
|
"loss": 0.5486, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 0.26280567049980164, |
|
"learning_rate": 6.122994652406417e-05, |
|
"loss": 0.3358, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 0.24697232246398926, |
|
"learning_rate": 5.8556149732620325e-05, |
|
"loss": 0.3871, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 0.22857435047626495, |
|
"learning_rate": 5.588235294117647e-05, |
|
"loss": 0.3822, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 0.2402905523777008, |
|
"learning_rate": 5.320855614973263e-05, |
|
"loss": 0.3009, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"grad_norm": 0.19268670678138733, |
|
"learning_rate": 5.0534759358288774e-05, |
|
"loss": 0.3916, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.2800000000000002, |
|
"grad_norm": 0.2288196086883545, |
|
"learning_rate": 4.786096256684492e-05, |
|
"loss": 0.338, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 0.2444332093000412, |
|
"learning_rate": 4.518716577540107e-05, |
|
"loss": 0.3971, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 0.26502081751823425, |
|
"learning_rate": 4.251336898395722e-05, |
|
"loss": 0.2905, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 2.392, |
|
"eval_loss": 0.45234861969947815, |
|
"eval_runtime": 967.7709, |
|
"eval_samples_per_second": 0.207, |
|
"eval_steps_per_second": 0.207, |
|
"step": 598 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 750, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 299, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.783168801781924e+18, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|