marimari-r2r-mlsum-clara-med
This model is a fine-tuned version of IIC/marimari-r2r-mlsum on the None dataset. It achieves the following results on the evaluation set:
- Loss: 3.9618
- Rouge1: 42.6764
- Rouge2: 24.4569
- Rougel: 37.0033
- Rougelsum: 37.1595
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5.6e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 30
Training results
Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum |
---|---|---|---|---|---|---|---|
No log | 1.0 | 190 | 2.3970 | 40.7426 | 23.212 | 35.7093 | 35.8437 |
No log | 2.0 | 380 | 2.3165 | 42.5676 | 24.6494 | 37.1225 | 37.2619 |
1.9699 | 3.0 | 570 | 2.4711 | 42.0346 | 23.7633 | 36.3472 | 36.4433 |
1.9699 | 4.0 | 760 | 2.7339 | 41.1717 | 22.8419 | 35.3263 | 35.4823 |
0.6485 | 5.0 | 950 | 2.9593 | 40.714 | 22.6931 | 34.8859 | 35.0647 |
0.6485 | 6.0 | 1140 | 3.1316 | 41.3218 | 23.2054 | 35.3103 | 35.5063 |
0.6485 | 7.0 | 1330 | 3.2542 | 41.2786 | 23.4853 | 35.8236 | 35.972 |
0.1529 | 8.0 | 1520 | 3.3470 | 41.2991 | 22.8385 | 35.0524 | 35.2153 |
0.1529 | 9.0 | 1710 | 3.4324 | 41.3838 | 23.1045 | 35.3472 | 35.5779 |
0.0719 | 10.0 | 1900 | 3.5187 | 42.0833 | 23.8538 | 36.3282 | 36.5294 |
0.0719 | 11.0 | 2090 | 3.5527 | 41.2993 | 23.0323 | 35.3116 | 35.4687 |
0.0719 | 12.0 | 2280 | 3.6624 | 41.6524 | 23.8925 | 35.9281 | 36.1012 |
0.0393 | 13.0 | 2470 | 3.6536 | 41.188 | 23.2066 | 35.371 | 35.5616 |
0.0393 | 14.0 | 2660 | 3.6656 | 40.8222 | 22.5651 | 35.0515 | 35.1399 |
0.0266 | 15.0 | 2850 | 3.7349 | 41.844 | 23.7839 | 36.102 | 36.3169 |
0.0266 | 16.0 | 3040 | 3.7254 | 41.5535 | 23.3996 | 35.9619 | 36.0981 |
0.0266 | 17.0 | 3230 | 3.7919 | 41.5683 | 23.2824 | 36.0855 | 36.2475 |
0.0151 | 18.0 | 3420 | 3.8152 | 42.1272 | 24.0548 | 36.5784 | 36.785 |
0.0151 | 19.0 | 3610 | 3.8213 | 41.9185 | 23.5975 | 36.1182 | 36.3194 |
0.0087 | 20.0 | 3800 | 3.8501 | 41.3409 | 23.0081 | 35.7662 | 35.9451 |
0.0087 | 21.0 | 3990 | 3.8690 | 41.9496 | 23.7032 | 36.0116 | 36.1843 |
0.0087 | 22.0 | 4180 | 3.8809 | 42.5366 | 24.6413 | 37.2644 | 37.459 |
0.0044 | 23.0 | 4370 | 3.8865 | 42.4346 | 24.2278 | 36.7284 | 36.8846 |
0.0044 | 24.0 | 4560 | 3.9044 | 42.9781 | 24.8423 | 37.3582 | 37.4807 |
0.0024 | 25.0 | 4750 | 3.9138 | 42.6738 | 24.4737 | 36.8959 | 37.0031 |
0.0024 | 26.0 | 4940 | 3.9361 | 42.5267 | 24.4155 | 36.8414 | 36.9915 |
0.0024 | 27.0 | 5130 | 3.9477 | 42.4844 | 24.5483 | 36.8857 | 37.0219 |
0.0013 | 28.0 | 5320 | 3.9561 | 42.7199 | 24.5977 | 37.1206 | 37.2374 |
0.0013 | 29.0 | 5510 | 3.9599 | 42.7088 | 24.4474 | 37.0513 | 37.1971 |
0.001 | 30.0 | 5700 | 3.9618 | 42.6764 | 24.4569 | 37.0033 | 37.1595 |
Framework versions
- Transformers 4.25.1
- Pytorch 1.13.0
- Datasets 2.8.0
- Tokenizers 0.12.1
- Downloads last month
- 17
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.