rh_qa_model / README.md
jpodivin's picture
End of training
59e14be verified
|
raw
history blame
3.78 kB
metadata
license: apache-2.0
base_model: distilbert-base-cased
tags:
  - generated_from_trainer
model-index:
  - name: rh_qa_model
    results: []

rh_qa_model

This model is a fine-tuned version of distilbert-base-cased on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 4.9421

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 50

Training results

Training Loss Epoch Step Validation Loss
No log 1.0 9 5.7606
No log 2.0 18 5.2134
No log 3.0 27 4.6264
No log 4.0 36 4.6808
No log 5.0 45 4.6072
No log 6.0 54 4.0322
No log 7.0 63 4.5964
No log 8.0 72 3.3412
No log 9.0 81 3.3570
No log 10.0 90 3.6391
No log 11.0 99 3.4018
No log 12.0 108 3.4077
No log 13.0 117 3.6908
No log 14.0 126 4.0146
No log 15.0 135 3.8527
No log 16.0 144 3.9721
No log 17.0 153 3.9417
No log 18.0 162 3.8664
No log 19.0 171 3.9009
No log 20.0 180 3.9753
No log 21.0 189 4.0617
No log 22.0 198 4.1928
No log 23.0 207 4.2910
No log 24.0 216 4.2968
No log 25.0 225 4.5358
No log 26.0 234 4.3997
No log 27.0 243 4.3952
No log 28.0 252 4.5261
No log 29.0 261 4.4806
No log 30.0 270 4.5944
No log 31.0 279 4.2314
No log 32.0 288 4.5624
No log 33.0 297 4.6074
No log 34.0 306 4.2452
No log 35.0 315 4.6662
No log 36.0 324 4.2687
No log 37.0 333 4.7763
No log 38.0 342 4.7474
No log 39.0 351 4.7573
No log 40.0 360 4.7578
No log 41.0 369 4.9119
No log 42.0 378 4.9752
No log 43.0 387 4.9227
No log 44.0 396 4.8575
No log 45.0 405 4.8022
No log 46.0 414 4.9180
No log 47.0 423 4.9006
No log 48.0 432 4.9202
No log 49.0 441 4.9380
No log 50.0 450 4.9421

Framework versions

  • Transformers 4.35.2
  • Pytorch 2.1.0+cu121
  • Datasets 2.16.1
  • Tokenizers 0.15.1