Writer12 / checkpoint-1158 /trainer_state.json
colinw2292's picture
Upload folder using huggingface_hub
c8b4b59 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 1158,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.2,
"learning_rate": 1.9137931034482762e-05,
"logits/chosen": -2.7039783000946045,
"logits/rejected": -2.699718713760376,
"logps/chosen": -178.0665283203125,
"logps/rejected": -202.87210083007812,
"loss": 0.65,
"rewards/accuracies": 0.798701286315918,
"rewards/chosen": -0.09282779693603516,
"rewards/margins": 0.09910193085670471,
"rewards/rejected": -0.19192971289157867,
"step": 77
},
{
"epoch": 0.4,
"learning_rate": 2.899232245681382e-05,
"logits/chosen": -2.7582526206970215,
"logits/rejected": -2.6498587131500244,
"logps/chosen": -206.910888671875,
"logps/rejected": -244.54962158203125,
"loss": 0.3215,
"rewards/accuracies": 0.9740259647369385,
"rewards/chosen": -2.6819589138031006,
"rewards/margins": 1.8207621574401855,
"rewards/rejected": -4.502721309661865,
"step": 154
},
{
"epoch": 0.6,
"learning_rate": 2.6833013435700576e-05,
"logits/chosen": -3.261249542236328,
"logits/rejected": -3.26186203956604,
"logps/chosen": -636.440673828125,
"logps/rejected": -844.8958740234375,
"loss": 0.1384,
"rewards/accuracies": 0.9675324559211731,
"rewards/chosen": -45.84988021850586,
"rewards/margins": 18.366718292236328,
"rewards/rejected": -64.21659088134766,
"step": 231
},
{
"epoch": 0.8,
"learning_rate": 2.4644913627639156e-05,
"logits/chosen": -3.799668788909912,
"logits/rejected": -3.877474308013916,
"logps/chosen": -1188.7479248046875,
"logps/rejected": -1513.479248046875,
"loss": 0.1428,
"rewards/accuracies": 0.9675324559211731,
"rewards/chosen": -100.75252532958984,
"rewards/margins": 30.6811580657959,
"rewards/rejected": -131.4336700439453,
"step": 308
},
{
"epoch": 1.0,
"learning_rate": 2.2428023032629558e-05,
"logits/chosen": -3.48297119140625,
"logits/rejected": -3.503549098968506,
"logps/chosen": -1211.92041015625,
"logps/rejected": -1585.6292724609375,
"loss": 0.3247,
"rewards/accuracies": 0.9545454382896423,
"rewards/chosen": -102.89164733886719,
"rewards/margins": 35.46805953979492,
"rewards/rejected": -138.3596954345703,
"step": 385
},
{
"epoch": 1.2,
"learning_rate": 2.0211132437619963e-05,
"logits/chosen": -3.9055564403533936,
"logits/rejected": -3.94558048248291,
"logps/chosen": -1156.05908203125,
"logps/rejected": -1524.2059326171875,
"loss": 0.2889,
"rewards/accuracies": 0.9545454382896423,
"rewards/chosen": -97.4389877319336,
"rewards/margins": 34.911155700683594,
"rewards/rejected": -132.3501434326172,
"step": 462
},
{
"epoch": 1.4,
"learning_rate": 1.7994241842610365e-05,
"logits/chosen": -4.063717365264893,
"logits/rejected": -4.160610675811768,
"logps/chosen": -1216.2169189453125,
"logps/rejected": -1579.79931640625,
"loss": 0.0666,
"rewards/accuracies": 0.9935064911842346,
"rewards/chosen": -103.59471130371094,
"rewards/margins": 34.364356994628906,
"rewards/rejected": -137.95909118652344,
"step": 539
},
{
"epoch": 1.6,
"learning_rate": 1.577735124760077e-05,
"logits/chosen": -3.7622299194335938,
"logits/rejected": -3.8654086589813232,
"logps/chosen": -1256.462158203125,
"logps/rejected": -1670.1668701171875,
"loss": 0.3068,
"rewards/accuracies": 0.9545454382896423,
"rewards/chosen": -107.87626647949219,
"rewards/margins": 38.972930908203125,
"rewards/rejected": -146.8491973876953,
"step": 616
},
{
"epoch": 1.8,
"learning_rate": 1.3589251439539348e-05,
"logits/chosen": -3.9780585765838623,
"logits/rejected": -4.0323262214660645,
"logps/chosen": -1205.580322265625,
"logps/rejected": -1609.4249267578125,
"loss": 0.0732,
"rewards/accuracies": 0.9675324559211731,
"rewards/chosen": -102.83528137207031,
"rewards/margins": 38.15076446533203,
"rewards/rejected": -140.9860382080078,
"step": 693
},
{
"epoch": 1.99,
"learning_rate": 1.137236084452975e-05,
"logits/chosen": -3.9996836185455322,
"logits/rejected": -4.038970947265625,
"logps/chosen": -1479.3536376953125,
"logps/rejected": -1951.1041259765625,
"loss": 0.2627,
"rewards/accuracies": 0.9935064911842346,
"rewards/chosen": -129.55506896972656,
"rewards/margins": 45.21900177001953,
"rewards/rejected": -174.77407836914062,
"step": 770
},
{
"epoch": 2.19,
"learning_rate": 9.155470249520153e-06,
"logits/chosen": -4.09077262878418,
"logits/rejected": -4.1275434494018555,
"logps/chosen": -1435.885986328125,
"logps/rejected": -1983.1124267578125,
"loss": 0.0968,
"rewards/accuracies": 0.9870129823684692,
"rewards/chosen": -125.91879272460938,
"rewards/margins": 52.251136779785156,
"rewards/rejected": -178.16995239257812,
"step": 847
},
{
"epoch": 2.39,
"learning_rate": 6.938579654510557e-06,
"logits/chosen": -4.158607482910156,
"logits/rejected": -4.210606098175049,
"logps/chosen": -1408.3914794921875,
"logps/rejected": -1875.4822998046875,
"loss": 0.0166,
"rewards/accuracies": 0.9805194735527039,
"rewards/chosen": -122.99951934814453,
"rewards/margins": 44.64362335205078,
"rewards/rejected": -167.64315795898438,
"step": 924
},
{
"epoch": 2.59,
"learning_rate": 4.72168905950096e-06,
"logits/chosen": -4.183628559112549,
"logits/rejected": -4.219183444976807,
"logps/chosen": -1309.217529296875,
"logps/rejected": -1762.13232421875,
"loss": 0.0602,
"rewards/accuracies": 0.9870129823684692,
"rewards/chosen": -113.17806243896484,
"rewards/margins": 43.657859802246094,
"rewards/rejected": -156.83592224121094,
"step": 1001
},
{
"epoch": 2.79,
"learning_rate": 2.5335892514395392e-06,
"logits/chosen": -4.1587114334106445,
"logits/rejected": -4.193912029266357,
"logps/chosen": -1347.27978515625,
"logps/rejected": -1834.71728515625,
"loss": 0.0365,
"rewards/accuracies": 0.9935064911842346,
"rewards/chosen": -116.40560150146484,
"rewards/margins": 46.630714416503906,
"rewards/rejected": -163.03631591796875,
"step": 1078
},
{
"epoch": 2.99,
"learning_rate": 3.166986564299424e-07,
"logits/chosen": -4.090827465057373,
"logits/rejected": -4.138672351837158,
"logps/chosen": -1456.274169921875,
"logps/rejected": -1961.9942626953125,
"loss": 0.0055,
"rewards/accuracies": 0.9935064911842346,
"rewards/chosen": -127.70254516601562,
"rewards/margins": 48.34178924560547,
"rewards/rejected": -176.04434204101562,
"step": 1155
}
],
"logging_steps": 77,
"max_steps": 1158,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"total_flos": 0.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}