mistral-dpo
This model is a fine-tuned version of TheBloke/Mistral-7B-v0.1-GPTQ on the None dataset. It achieves the following results on the evaluation set:
- Loss: 0.5603
- Rewards/chosen: -12.5467
- Rewards/rejected: -28.4037
- Rewards/accuracies: 0.8571
- Rewards/margins: 15.8571
- Logps/rejected: -411.7001
- Logps/chosen: -215.4742
- Logits/rejected: -0.7509
- Logits/chosen: -0.7707
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 1
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 2
- training_steps: 250
- mixed_precision_training: Native AMP
Training results
Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
---|---|---|---|---|---|---|---|---|---|---|---|
0.6785 | 0.02 | 10 | 0.6291 | -0.0030 | -0.1321 | 0.875 | 0.1291 | -128.9836 | -90.0372 | -2.3988 | -2.3489 |
0.5661 | 0.04 | 20 | 0.4421 | 0.0008 | -0.6608 | 0.875 | 0.6616 | -134.2708 | -89.9997 | -2.3613 | -2.3042 |
0.3257 | 0.06 | 30 | 0.3584 | -0.7139 | -2.3035 | 0.8393 | 1.5897 | -150.6985 | -97.1463 | -2.2995 | -2.2546 |
0.3563 | 0.08 | 40 | 0.5522 | -3.0636 | -6.7067 | 0.8214 | 3.6431 | -194.7305 | -120.6441 | -2.1396 | -2.0849 |
0.0013 | 0.1 | 50 | 1.3365 | -8.4317 | -16.1649 | 0.8036 | 7.7332 | -289.3120 | -174.3246 | -1.8243 | -1.7710 |
0.0277 | 0.12 | 60 | 2.4224 | -14.8512 | -25.9570 | 0.8214 | 11.1059 | -387.2331 | -238.5192 | -1.5464 | -1.4950 |
1.5742 | 0.14 | 70 | 3.1075 | -17.8751 | -29.6755 | 0.8214 | 11.8004 | -424.4176 | -268.7585 | -1.4071 | -1.3681 |
14.1036 | 0.16 | 80 | 3.6238 | -20.4205 | -32.7881 | 0.8214 | 12.3675 | -455.5435 | -294.2129 | -1.3237 | -1.2729 |
8.531 | 0.18 | 90 | 3.7123 | -21.7863 | -36.0729 | 0.8214 | 14.2866 | -488.3922 | -307.8707 | -1.2975 | -1.2388 |
4.6429 | 0.2 | 100 | 2.0394 | -16.6472 | -29.8508 | 0.8393 | 13.2036 | -426.1712 | -256.4797 | -1.3189 | -1.2784 |
0.0565 | 0.22 | 110 | 1.6331 | -14.8501 | -27.2015 | 0.8393 | 12.3514 | -399.6779 | -238.5090 | -1.2425 | -1.2118 |
0.0056 | 0.24 | 120 | 1.4774 | -15.0784 | -28.0012 | 0.8214 | 12.9228 | -407.6750 | -240.7916 | -1.0819 | -1.0579 |
0.0804 | 0.26 | 130 | 1.5398 | -16.7630 | -30.6346 | 0.8393 | 13.8716 | -434.0091 | -257.6378 | -1.0054 | -0.9846 |
0.0001 | 0.28 | 140 | 1.5159 | -17.9940 | -33.3459 | 0.8393 | 15.3520 | -461.1225 | -269.9474 | -0.8887 | -0.8844 |
0.0 | 0.3 | 150 | 1.5062 | -18.4614 | -34.3481 | 0.8393 | 15.8868 | -471.1445 | -274.6213 | -0.8496 | -0.8503 |
0.0 | 0.32 | 160 | 1.5035 | -18.6474 | -34.7165 | 0.8393 | 16.0692 | -474.8286 | -276.4815 | -0.8343 | -0.8367 |
4.2123 | 0.34 | 170 | 1.2949 | -17.3471 | -32.6721 | 0.8571 | 15.3250 | -454.3839 | -263.4789 | -0.8672 | -0.8661 |
2.13 | 0.36 | 180 | 0.9892 | -15.2178 | -30.1177 | 0.8571 | 14.8999 | -428.8398 | -242.1859 | -0.8992 | -0.9047 |
2.0146 | 0.38 | 190 | 0.8365 | -13.9461 | -28.5983 | 0.8571 | 14.6522 | -413.6459 | -229.4683 | -0.9104 | -0.9224 |
0.0706 | 0.4 | 200 | 0.7897 | -14.5198 | -29.8989 | 0.8571 | 15.3791 | -426.6525 | -235.2058 | -0.8064 | -0.8224 |
5.2517 | 0.42 | 210 | 0.6621 | -13.7049 | -29.2354 | 0.8571 | 15.5305 | -420.0170 | -227.0569 | -0.7981 | -0.8124 |
0.0499 | 0.44 | 220 | 0.5752 | -12.8733 | -28.5299 | 0.8571 | 15.6566 | -412.9616 | -218.7403 | -0.7801 | -0.7990 |
0.5779 | 0.46 | 230 | 0.5611 | -12.6633 | -28.3836 | 0.8571 | 15.7203 | -411.4988 | -216.6405 | -0.7789 | -0.7975 |
0.0322 | 0.48 | 240 | 0.5624 | -12.6348 | -28.4766 | 0.8571 | 15.8418 | -412.4289 | -216.3556 | -0.7696 | -0.7878 |
0.1347 | 0.5 | 250 | 0.5603 | -12.5467 | -28.4037 | 0.8571 | 15.8571 | -411.7001 | -215.4742 | -0.7509 | -0.7707 |
Framework versions
- PEFT 0.7.1
- Transformers 4.36.2
- Pytorch 2.0.1+cu118
- Datasets 2.15.0
- Tokenizers 0.15.0
- Downloads last month
- 2
Model tree for Belred/mistral-dpo
Base model
mistralai/Mistral-7B-v0.1
Quantized
TheBloke/Mistral-7B-v0.1-GPTQ