|
{ |
|
"best_metric": 2.269778251647949, |
|
"best_model_checkpoint": "../../saves/Qwen1.5-7B-Chat/lora/sft/checkpoint-1200", |
|
"epoch": 7.111111111111111, |
|
"eval_steps": 400, |
|
"global_step": 1200, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 2.701119899749756, |
|
"learning_rate": 2.5e-05, |
|
"loss": 4.3102, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 3.890558958053589, |
|
"learning_rate": 5e-05, |
|
"loss": 3.8998, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.3287036418914795, |
|
"learning_rate": 4.999552306674344e-05, |
|
"loss": 3.241, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.6666322946548462, |
|
"learning_rate": 4.998209387040829e-05, |
|
"loss": 2.677, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.6145800352096558, |
|
"learning_rate": 4.9959717220723784e-05, |
|
"loss": 2.6172, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.5234311819076538, |
|
"learning_rate": 4.9928401131991305e-05, |
|
"loss": 2.5621, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.6433086395263672, |
|
"learning_rate": 4.9888156820213974e-05, |
|
"loss": 2.587, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.5955637693405151, |
|
"learning_rate": 4.9838998699079625e-05, |
|
"loss": 2.6315, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.5844753980636597, |
|
"learning_rate": 4.9780944374798435e-05, |
|
"loss": 2.5056, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.6106901168823242, |
|
"learning_rate": 4.971401463979721e-05, |
|
"loss": 2.5099, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.7099446654319763, |
|
"learning_rate": 4.963823346527248e-05, |
|
"loss": 2.4891, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.6092314720153809, |
|
"learning_rate": 4.9553627992605066e-05, |
|
"loss": 2.546, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.5691718459129333, |
|
"learning_rate": 4.946022852363932e-05, |
|
"loss": 2.4562, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.7633320689201355, |
|
"learning_rate": 4.9358068509830334e-05, |
|
"loss": 2.5554, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.7102218866348267, |
|
"learning_rate": 4.924718454026318e-05, |
|
"loss": 2.5479, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.7560188174247742, |
|
"learning_rate": 4.912761632854833e-05, |
|
"loss": 2.4747, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 0.6717157959938049, |
|
"learning_rate": 4.8999406698598074e-05, |
|
"loss": 2.5173, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 0.6935433745384216, |
|
"learning_rate": 4.886260156928888e-05, |
|
"loss": 2.4288, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 1.0115560293197632, |
|
"learning_rate": 4.8717249938015415e-05, |
|
"loss": 2.5331, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.7541220784187317, |
|
"learning_rate": 4.856340386314182e-05, |
|
"loss": 2.4577, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.8513575196266174, |
|
"learning_rate": 4.840111844535682e-05, |
|
"loss": 2.496, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.7479698061943054, |
|
"learning_rate": 4.8230451807939135e-05, |
|
"loss": 2.4142, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 0.7310675978660583, |
|
"learning_rate": 4.8051465075940336e-05, |
|
"loss": 2.4913, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 0.8907930850982666, |
|
"learning_rate": 4.786422235429269e-05, |
|
"loss": 2.3697, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.9348196387290955, |
|
"learning_rate": 4.766879070484956e-05, |
|
"loss": 2.4631, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 0.9181344509124756, |
|
"learning_rate": 4.746524012236706e-05, |
|
"loss": 2.4411, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.7747082710266113, |
|
"learning_rate": 4.725364350943492e-05, |
|
"loss": 2.3592, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 0.9620688557624817, |
|
"learning_rate": 4.703407665036622e-05, |
|
"loss": 2.3433, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 1.016157865524292, |
|
"learning_rate": 4.680661818405485e-05, |
|
"loss": 2.3396, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 0.8236090540885925, |
|
"learning_rate": 4.657134957581057e-05, |
|
"loss": 2.3637, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 0.9651027321815491, |
|
"learning_rate": 4.6328355088181915e-05, |
|
"loss": 2.3831, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 0.7949943542480469, |
|
"learning_rate": 4.607772175077711e-05, |
|
"loss": 2.4422, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 0.9031904935836792, |
|
"learning_rate": 4.581953932909403e-05, |
|
"loss": 2.4595, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 0.859309732913971, |
|
"learning_rate": 4.555390029237026e-05, |
|
"loss": 2.3967, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"grad_norm": 1.1613001823425293, |
|
"learning_rate": 4.528089978046481e-05, |
|
"loss": 2.3132, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"grad_norm": 0.9616575837135315, |
|
"learning_rate": 4.500063556978337e-05, |
|
"loss": 2.4425, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 0.7751485109329224, |
|
"learning_rate": 4.471320803825915e-05, |
|
"loss": 2.3591, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 1.3079149723052979, |
|
"learning_rate": 4.441872012940214e-05, |
|
"loss": 2.3323, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 0.8369519710540771, |
|
"learning_rate": 4.4117277315429366e-05, |
|
"loss": 2.5005, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 1.0954837799072266, |
|
"learning_rate": 4.380898755948953e-05, |
|
"loss": 2.3058, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"eval_loss": 2.299175500869751, |
|
"eval_runtime": 74.4017, |
|
"eval_samples_per_second": 8.064, |
|
"eval_steps_per_second": 4.032, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 1.0771162509918213, |
|
"learning_rate": 4.349396127699552e-05, |
|
"loss": 2.4145, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 0.7209674119949341, |
|
"learning_rate": 4.3172311296078595e-05, |
|
"loss": 2.3519, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 1.0076887607574463, |
|
"learning_rate": 4.284415281717847e-05, |
|
"loss": 2.3699, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"grad_norm": 0.9578609466552734, |
|
"learning_rate": 4.250960337178377e-05, |
|
"loss": 2.3923, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"grad_norm": 0.9023878574371338, |
|
"learning_rate": 4.216878278033753e-05, |
|
"loss": 2.3618, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"grad_norm": 1.1171501874923706, |
|
"learning_rate": 4.1821813109322974e-05, |
|
"loss": 2.4161, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 1.2200597524642944, |
|
"learning_rate": 4.1468818627544845e-05, |
|
"loss": 2.3592, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"grad_norm": 0.9814207553863525, |
|
"learning_rate": 4.1109925761621925e-05, |
|
"loss": 2.4143, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 1.2403128147125244, |
|
"learning_rate": 4.0745263050706784e-05, |
|
"loss": 2.3558, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 0.9680503606796265, |
|
"learning_rate": 4.037496110044884e-05, |
|
"loss": 2.3103, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"grad_norm": 0.9476011991500854, |
|
"learning_rate": 3.999915253621739e-05, |
|
"loss": 2.369, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"grad_norm": 0.9985129237174988, |
|
"learning_rate": 3.961797195560118e-05, |
|
"loss": 2.2797, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"grad_norm": 1.1295677423477173, |
|
"learning_rate": 3.9231555880201655e-05, |
|
"loss": 2.3331, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"grad_norm": 0.854465126991272, |
|
"learning_rate": 3.8840042706737114e-05, |
|
"loss": 2.2353, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"grad_norm": 0.9416671991348267, |
|
"learning_rate": 3.8443572657475304e-05, |
|
"loss": 2.285, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"grad_norm": 1.1753567457199097, |
|
"learning_rate": 3.804228773001212e-05, |
|
"loss": 2.3359, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"grad_norm": 1.0389409065246582, |
|
"learning_rate": 3.7636331646414524e-05, |
|
"loss": 2.3567, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"grad_norm": 1.197994589805603, |
|
"learning_rate": 3.7225849801745835e-05, |
|
"loss": 2.3137, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"grad_norm": 1.1205960512161255, |
|
"learning_rate": 3.6810989211991774e-05, |
|
"loss": 2.3268, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"grad_norm": 1.1302969455718994, |
|
"learning_rate": 3.639189846140604e-05, |
|
"loss": 2.4254, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"grad_norm": 1.1475387811660767, |
|
"learning_rate": 3.596872764929413e-05, |
|
"loss": 2.3282, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"grad_norm": 1.2079755067825317, |
|
"learning_rate": 3.55416283362546e-05, |
|
"loss": 2.2859, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"grad_norm": 1.274043321609497, |
|
"learning_rate": 3.511075348989692e-05, |
|
"loss": 2.3619, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"grad_norm": 1.3108967542648315, |
|
"learning_rate": 3.4676257430055434e-05, |
|
"loss": 2.202, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"grad_norm": 1.1571152210235596, |
|
"learning_rate": 3.4238295773518924e-05, |
|
"loss": 2.3114, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"grad_norm": 1.2922154664993286, |
|
"learning_rate": 3.379702537829583e-05, |
|
"loss": 2.3175, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"grad_norm": 0.8678541779518127, |
|
"learning_rate": 3.335260428743475e-05, |
|
"loss": 2.4348, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"grad_norm": 1.2109763622283936, |
|
"learning_rate": 3.29051916724206e-05, |
|
"loss": 2.3685, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"grad_norm": 1.21372389793396, |
|
"learning_rate": 3.2454947776166636e-05, |
|
"loss": 2.1795, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"grad_norm": 1.2954655885696411, |
|
"learning_rate": 3.200203385562268e-05, |
|
"loss": 2.2299, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"grad_norm": 1.4054641723632812, |
|
"learning_rate": 3.154661212402017e-05, |
|
"loss": 2.3114, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"grad_norm": 1.2466893196105957, |
|
"learning_rate": 3.10888456927748e-05, |
|
"loss": 2.2702, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"grad_norm": 1.0978355407714844, |
|
"learning_rate": 3.0628898513067353e-05, |
|
"loss": 2.2537, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"grad_norm": 1.3766223192214966, |
|
"learning_rate": 3.0166935317123823e-05, |
|
"loss": 2.3131, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"grad_norm": 1.152358889579773, |
|
"learning_rate": 2.9703121559215845e-05, |
|
"loss": 2.38, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"grad_norm": 1.3584524393081665, |
|
"learning_rate": 2.923762335640242e-05, |
|
"loss": 2.3432, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 4.56, |
|
"grad_norm": 1.3569154739379883, |
|
"learning_rate": 2.8770607429034352e-05, |
|
"loss": 2.2632, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"grad_norm": 1.4426939487457275, |
|
"learning_rate": 2.8302241041042565e-05, |
|
"loss": 2.2501, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"grad_norm": 1.2213878631591797, |
|
"learning_rate": 2.783269194003175e-05, |
|
"loss": 2.3787, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"grad_norm": 1.5808135271072388, |
|
"learning_rate": 2.7362128297200785e-05, |
|
"loss": 2.3647, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"eval_loss": 2.2738687992095947, |
|
"eval_runtime": 80.6321, |
|
"eval_samples_per_second": 7.441, |
|
"eval_steps_per_second": 3.721, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"grad_norm": 1.2698041200637817, |
|
"learning_rate": 2.6890718647111422e-05, |
|
"loss": 2.2942, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"grad_norm": 1.2770848274230957, |
|
"learning_rate": 2.6418631827326857e-05, |
|
"loss": 2.347, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 4.92, |
|
"grad_norm": 1.2166974544525146, |
|
"learning_rate": 2.5946036917941762e-05, |
|
"loss": 2.2969, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"grad_norm": 1.4275230169296265, |
|
"learning_rate": 2.5473103181025476e-05, |
|
"loss": 2.2556, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"grad_norm": 1.2852686643600464, |
|
"learning_rate": 2.5e-05, |
|
"loss": 2.2781, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"grad_norm": 1.6887468099594116, |
|
"learning_rate": 2.4526896818974533e-05, |
|
"loss": 2.2843, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"grad_norm": 1.328687310218811, |
|
"learning_rate": 2.4053963082058244e-05, |
|
"loss": 2.229, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"grad_norm": 1.3591057062149048, |
|
"learning_rate": 2.3581368172673152e-05, |
|
"loss": 2.2851, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 5.27, |
|
"grad_norm": 1.3400282859802246, |
|
"learning_rate": 2.310928135288859e-05, |
|
"loss": 2.18, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 5.33, |
|
"grad_norm": 1.3457157611846924, |
|
"learning_rate": 2.263787170279922e-05, |
|
"loss": 2.2501, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 5.39, |
|
"grad_norm": 1.5690137147903442, |
|
"learning_rate": 2.2167308059968254e-05, |
|
"loss": 2.2682, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 5.45, |
|
"grad_norm": 1.429304599761963, |
|
"learning_rate": 2.1697758958957448e-05, |
|
"loss": 2.3144, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 5.51, |
|
"grad_norm": 1.4117275476455688, |
|
"learning_rate": 2.1229392570965657e-05, |
|
"loss": 2.2955, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 5.57, |
|
"grad_norm": 1.7679541110992432, |
|
"learning_rate": 2.0762376643597582e-05, |
|
"loss": 2.2117, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 5.63, |
|
"grad_norm": 1.4710173606872559, |
|
"learning_rate": 2.029687844078416e-05, |
|
"loss": 2.3107, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"grad_norm": 1.2774254083633423, |
|
"learning_rate": 1.9833064682876176e-05, |
|
"loss": 2.2242, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 5.75, |
|
"grad_norm": 1.5879724025726318, |
|
"learning_rate": 1.937110148693265e-05, |
|
"loss": 2.2299, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 5.81, |
|
"grad_norm": 1.3109989166259766, |
|
"learning_rate": 1.8911154307225203e-05, |
|
"loss": 2.3555, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 5.87, |
|
"grad_norm": 1.6713744401931763, |
|
"learning_rate": 1.8453387875979834e-05, |
|
"loss": 2.2603, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 5.93, |
|
"grad_norm": 1.6011615991592407, |
|
"learning_rate": 1.7997966144377325e-05, |
|
"loss": 2.2731, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 5.99, |
|
"grad_norm": 1.4683603048324585, |
|
"learning_rate": 1.754505222383337e-05, |
|
"loss": 2.2876, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"grad_norm": 1.3678452968597412, |
|
"learning_rate": 1.70948083275794e-05, |
|
"loss": 2.3123, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 6.1, |
|
"grad_norm": 1.4179573059082031, |
|
"learning_rate": 1.6647395712565256e-05, |
|
"loss": 2.1601, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 6.16, |
|
"grad_norm": 1.4710493087768555, |
|
"learning_rate": 1.6202974621704175e-05, |
|
"loss": 2.1918, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 6.22, |
|
"grad_norm": 1.5775597095489502, |
|
"learning_rate": 1.576170422648108e-05, |
|
"loss": 2.3438, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 6.28, |
|
"grad_norm": 1.7641725540161133, |
|
"learning_rate": 1.5323742569944572e-05, |
|
"loss": 2.1773, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 6.34, |
|
"grad_norm": 1.5010628700256348, |
|
"learning_rate": 1.4889246510103077e-05, |
|
"loss": 2.2014, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"grad_norm": 1.5155980587005615, |
|
"learning_rate": 1.4458371663745401e-05, |
|
"loss": 2.2616, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"grad_norm": 1.337975025177002, |
|
"learning_rate": 1.4031272350705871e-05, |
|
"loss": 2.2648, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 6.52, |
|
"grad_norm": 1.788812518119812, |
|
"learning_rate": 1.3608101538593965e-05, |
|
"loss": 2.2841, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 6.58, |
|
"grad_norm": 1.6673487424850464, |
|
"learning_rate": 1.3189010788008233e-05, |
|
"loss": 2.3188, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 6.64, |
|
"grad_norm": 1.4954336881637573, |
|
"learning_rate": 1.277415019825417e-05, |
|
"loss": 2.2268, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 6.7, |
|
"grad_norm": 1.5983794927597046, |
|
"learning_rate": 1.2363668353585487e-05, |
|
"loss": 2.2214, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 6.76, |
|
"grad_norm": 1.6327165365219116, |
|
"learning_rate": 1.195771226998789e-05, |
|
"loss": 2.2109, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"grad_norm": 1.7117351293563843, |
|
"learning_rate": 1.1556427342524698e-05, |
|
"loss": 2.3153, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 6.87, |
|
"grad_norm": 1.8630892038345337, |
|
"learning_rate": 1.1159957293262888e-05, |
|
"loss": 2.216, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 6.93, |
|
"grad_norm": 1.6396968364715576, |
|
"learning_rate": 1.0768444119798357e-05, |
|
"loss": 2.2552, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 6.99, |
|
"grad_norm": 1.5091443061828613, |
|
"learning_rate": 1.0382028044398822e-05, |
|
"loss": 2.2435, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 7.05, |
|
"grad_norm": 1.4590834379196167, |
|
"learning_rate": 1.0000847463782615e-05, |
|
"loss": 2.2492, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 7.11, |
|
"grad_norm": 1.8363531827926636, |
|
"learning_rate": 9.625038899551161e-06, |
|
"loss": 2.1918, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 7.11, |
|
"eval_loss": 2.269778251647949, |
|
"eval_runtime": 74.1482, |
|
"eval_samples_per_second": 8.092, |
|
"eval_steps_per_second": 4.046, |
|
"step": 1200 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 1680, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 400, |
|
"total_flos": 1.1235201502895145e+18, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|