File size: 342 Bytes
fb5a8b8
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
This model is a fine-tune checkpoint of DistilBERT-base-uncased, fine-tuned on SST-2. This model reaches an accuracy of 91.3 on the dev set (for comparison, Bert bert-base-uncased version reaches an accuracy of 92.7).

Fine-tuning hyper-parameters
learning_rate = 1e-5
batch_size = 32
warmup = 600
max_seq_length = 128
num_train_epochs = 3.0