File size: 1,038 Bytes
c82fbcd 14edb25 bafdfc4 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 c82fbcd 14edb25 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 |
---
metrics:
- accuracy
- precision
pipeline_tag: token-classification
---
tokenclass-wnut
This model is a fine-tuned version of distilbert-base-uncased on the wnut_17 dataset. It achieves the following results on the evaluation set:
Loss: 0.2858
Precision: 0.4846
Recall: 0.2632
F1: 0.3411
Accuracy: 0.9386
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
learning_rate: 2e-05
train_batch_size: 16
eval_batch_size: 16
seed: 42
optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
lr_scheduler_type: linear
num_epochs: 2
Training results
Training Loss Epoch Step Validation Loss Precision Recall F1 Accuracy
No log 1.0 213 0.2976 0.3873 0.1974 0.2615 0.9352
No log 2.0 426 0.2858 0.4846 0.2632 0.3411 0.9386
Framework versions
Transformers 4.20.1
Pytorch 1.11.0+cpu
Datasets 2.1.0
Tokenizers 0.12.1 |