ironrock's picture
Upload folder using huggingface_hub
7eb44e2 verified
raw
history blame
7.13 kB
{
"best_metric": 0.4401554763317108,
"best_model_checkpoint": "./mistral/19-04-24-Weni-WeniGPT-Agents-Mistral-1.0.6-SFT-1.0.3-DPO_Experiment on DPO with other hyperparameters and best SFT model of WeniGPT-2_max_steps-366_batch_4_2024-04-19_ppid_9/checkpoint-90",
"epoch": 1.4634146341463414,
"eval_steps": 30,
"global_step": 90,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.16,
"grad_norm": 29.063692092895508,
"learning_rate": 3.181818181818182e-06,
"logits/chosen": -1.7455532550811768,
"logits/rejected": -1.7793405055999756,
"logps/chosen": -54.72136306762695,
"logps/rejected": -97.28812408447266,
"loss": 0.6924,
"rewards/accuracies": 0.15000000596046448,
"rewards/chosen": 0.003122234484180808,
"rewards/margins": 0.0015356539515778422,
"rewards/rejected": 0.0015865802997723222,
"step": 10
},
{
"epoch": 0.33,
"grad_norm": 15.157337188720703,
"learning_rate": 4.9154929577464795e-06,
"logits/chosen": -1.8641210794448853,
"logits/rejected": -1.8705828189849854,
"logps/chosen": -40.3618049621582,
"logps/rejected": -55.96257400512695,
"loss": 0.6737,
"rewards/accuracies": 0.32499998807907104,
"rewards/chosen": 0.04275550693273544,
"rewards/margins": 0.040880247950553894,
"rewards/rejected": 0.0018752537434920669,
"step": 20
},
{
"epoch": 0.49,
"grad_norm": 13.148392677307129,
"learning_rate": 4.774647887323944e-06,
"logits/chosen": -1.8284633159637451,
"logits/rejected": -1.8519256114959717,
"logps/chosen": -44.69489288330078,
"logps/rejected": -76.79936981201172,
"loss": 0.6371,
"rewards/accuracies": 0.42500001192092896,
"rewards/chosen": 0.12410124391317368,
"rewards/margins": 0.1248731017112732,
"rewards/rejected": -0.0007718756678514183,
"step": 30
},
{
"epoch": 0.49,
"eval_logits/chosen": -1.7379764318466187,
"eval_logits/rejected": -1.7726188898086548,
"eval_logps/chosen": -50.49373245239258,
"eval_logps/rejected": -83.54340362548828,
"eval_loss": 0.5864880681037903,
"eval_rewards/accuracies": 0.4642857015132904,
"eval_rewards/chosen": 0.24716782569885254,
"eval_rewards/margins": 0.25581416487693787,
"eval_rewards/rejected": -0.008646338246762753,
"eval_runtime": 8.1282,
"eval_samples_per_second": 3.445,
"eval_steps_per_second": 1.722,
"step": 30
},
{
"epoch": 0.65,
"grad_norm": 24.143388748168945,
"learning_rate": 4.633802816901409e-06,
"logits/chosen": -1.8080145120620728,
"logits/rejected": -1.8319604396820068,
"logps/chosen": -46.848201751708984,
"logps/rejected": -81.42877960205078,
"loss": 0.5649,
"rewards/accuracies": 0.42500001192092896,
"rewards/chosen": 0.3360167443752289,
"rewards/margins": 0.34645381569862366,
"rewards/rejected": -0.010437069460749626,
"step": 40
},
{
"epoch": 0.81,
"grad_norm": 0.0,
"learning_rate": 4.492957746478874e-06,
"logits/chosen": -1.8638660907745361,
"logits/rejected": -1.8754875659942627,
"logps/chosen": -22.06886863708496,
"logps/rejected": -41.17422103881836,
"loss": 0.6377,
"rewards/accuracies": 0.20000000298023224,
"rewards/chosen": 0.25195080041885376,
"rewards/margins": 0.16642150282859802,
"rewards/rejected": 0.08552931249141693,
"step": 50
},
{
"epoch": 0.98,
"grad_norm": 7.51575231552124,
"learning_rate": 4.352112676056338e-06,
"logits/chosen": -1.8215672969818115,
"logits/rejected": -1.8419084548950195,
"logps/chosen": -44.14409255981445,
"logps/rejected": -77.38871002197266,
"loss": 0.5496,
"rewards/accuracies": 0.3499999940395355,
"rewards/chosen": 0.4203863739967346,
"rewards/margins": 0.458610475063324,
"rewards/rejected": -0.038224123418331146,
"step": 60
},
{
"epoch": 0.98,
"eval_logits/chosen": -1.7419661283493042,
"eval_logits/rejected": -1.7773857116699219,
"eval_logps/chosen": -49.36268997192383,
"eval_logps/rejected": -83.60606384277344,
"eval_loss": 0.4963766038417816,
"eval_rewards/accuracies": 0.4642857015132904,
"eval_rewards/chosen": 0.5864797234535217,
"eval_rewards/margins": 0.613922655582428,
"eval_rewards/rejected": -0.02744293212890625,
"eval_runtime": 8.1315,
"eval_samples_per_second": 3.443,
"eval_steps_per_second": 1.722,
"step": 60
},
{
"epoch": 1.14,
"grad_norm": 3.454394817352295,
"learning_rate": 4.211267605633803e-06,
"logits/chosen": -1.7844343185424805,
"logits/rejected": -1.8126220703125,
"logps/chosen": -39.737998962402344,
"logps/rejected": -78.76033782958984,
"loss": 0.4925,
"rewards/accuracies": 0.4000000059604645,
"rewards/chosen": 0.6417733430862427,
"rewards/margins": 0.7573493123054504,
"rewards/rejected": -0.11557593196630478,
"step": 70
},
{
"epoch": 1.3,
"grad_norm": 14.345088005065918,
"learning_rate": 4.070422535211268e-06,
"logits/chosen": -1.822148084640503,
"logits/rejected": -1.8463541269302368,
"logps/chosen": -35.735836029052734,
"logps/rejected": -75.4683837890625,
"loss": 0.4965,
"rewards/accuracies": 0.375,
"rewards/chosen": 0.6439998745918274,
"rewards/margins": 0.8037986755371094,
"rewards/rejected": -0.1597987860441208,
"step": 80
},
{
"epoch": 1.46,
"grad_norm": 12.988151550292969,
"learning_rate": 3.9295774647887325e-06,
"logits/chosen": -1.858496069908142,
"logits/rejected": -1.8683990240097046,
"logps/chosen": -30.166736602783203,
"logps/rejected": -50.970542907714844,
"loss": 0.5185,
"rewards/accuracies": 0.30000001192092896,
"rewards/chosen": 0.6545813679695129,
"rewards/margins": 0.7556509971618652,
"rewards/rejected": -0.10106959193944931,
"step": 90
},
{
"epoch": 1.46,
"eval_logits/chosen": -1.7461166381835938,
"eval_logits/rejected": -1.7827314138412476,
"eval_logps/chosen": -47.9538688659668,
"eval_logps/rejected": -83.84149169921875,
"eval_loss": 0.4401554763317108,
"eval_rewards/accuracies": 0.4642857015132904,
"eval_rewards/chosen": 1.0091263055801392,
"eval_rewards/margins": 1.1071958541870117,
"eval_rewards/rejected": -0.09806957095861435,
"eval_runtime": 8.1286,
"eval_samples_per_second": 3.445,
"eval_steps_per_second": 1.722,
"step": 90
}
],
"logging_steps": 10,
"max_steps": 366,
"num_input_tokens_seen": 0,
"num_train_epochs": 6,
"save_steps": 90,
"total_flos": 0.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}