tm21cy commited on
Commit
9c06c34
·
verified ·
1 Parent(s): 86f6e91

End of training

Browse files
Files changed (4) hide show
  1. README.md +10 -9
  2. config.json +24 -15
  3. model.safetensors +2 -2
  4. training_args.bin +1 -1
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: apache-2.0
3
- base_model: distilbert-base-uncased
4
  tags:
5
  - generated_from_trainer
6
  metrics:
@@ -15,10 +15,10 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # results
17
 
18
- This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.8835
21
- - Accuracy: 0.6862
22
 
23
  ## Model description
24
 
@@ -37,21 +37,22 @@ More information needed
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
- - learning_rate: 5e-05
41
  - train_batch_size: 16
42
  - eval_batch_size: 16
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
- - num_epochs: 3
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
51
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
52
- | No log | 1.0 | 179 | 0.6302 | 0.6440 |
53
- | No log | 2.0 | 358 | 0.6395 | 0.6712 |
54
- | 0.4692 | 3.0 | 537 | 0.8529 | 0.6764 |
 
55
 
56
 
57
  ### Framework versions
 
1
  ---
2
  license: apache-2.0
3
+ base_model: albert-base-v2
4
  tags:
5
  - generated_from_trainer
6
  metrics:
 
15
 
16
  # results
17
 
18
+ This model is a fine-tuned version of [albert-base-v2](https://huggingface.co/albert-base-v2) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.7940
21
+ - Accuracy: 0.6556
22
 
23
  ## Model description
24
 
 
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
+ - learning_rate: 7.45e-06
41
  - train_batch_size: 16
42
  - eval_batch_size: 16
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
+ - num_epochs: 4
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
51
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
52
+ | No log | 1.0 | 179 | 0.6503 | 0.6492 |
53
+ | No log | 2.0 | 358 | 0.7322 | 0.6565 |
54
+ | 0.4518 | 3.0 | 537 | 0.7242 | 0.6649 |
55
+ | 0.4518 | 4.0 | 716 | 0.7997 | 0.6586 |
56
 
57
 
58
  ### Framework versions
config.json CHANGED
@@ -1,25 +1,34 @@
1
  {
2
- "_name_or_path": "distilbert-base-uncased",
3
- "activation": "gelu",
4
  "architectures": [
5
- "DistilBertForSequenceClassification"
6
  ],
7
- "attention_dropout": 0.1,
8
- "dim": 768,
9
- "dropout": 0.1,
10
- "hidden_dim": 3072,
 
 
 
 
 
 
11
  "initializer_range": 0.02,
 
 
 
12
  "max_position_embeddings": 512,
13
- "model_type": "distilbert",
14
- "n_heads": 12,
15
- "n_layers": 6,
 
 
 
16
  "pad_token_id": 0,
 
17
  "problem_type": "single_label_classification",
18
- "qa_dropout": 0.1,
19
- "seq_classif_dropout": 0.2,
20
- "sinusoidal_pos_embds": false,
21
- "tie_weights_": true,
22
  "torch_dtype": "float32",
23
  "transformers_version": "4.38.2",
24
- "vocab_size": 30522
 
25
  }
 
1
  {
2
+ "_name_or_path": "albert-base-v2",
 
3
  "architectures": [
4
+ "AlbertForSequenceClassification"
5
  ],
6
+ "attention_probs_dropout_prob": 0,
7
+ "bos_token_id": 2,
8
+ "classifier_dropout_prob": 0.1,
9
+ "down_scale_factor": 1,
10
+ "embedding_size": 128,
11
+ "eos_token_id": 3,
12
+ "gap_size": 0,
13
+ "hidden_act": "gelu_new",
14
+ "hidden_dropout_prob": 0,
15
+ "hidden_size": 768,
16
  "initializer_range": 0.02,
17
+ "inner_group_num": 1,
18
+ "intermediate_size": 3072,
19
+ "layer_norm_eps": 1e-12,
20
  "max_position_embeddings": 512,
21
+ "model_type": "albert",
22
+ "net_structure_type": 0,
23
+ "num_attention_heads": 12,
24
+ "num_hidden_groups": 1,
25
+ "num_hidden_layers": 12,
26
+ "num_memory_blocks": 0,
27
  "pad_token_id": 0,
28
+ "position_embedding_type": "absolute",
29
  "problem_type": "single_label_classification",
 
 
 
 
30
  "torch_dtype": "float32",
31
  "transformers_version": "4.38.2",
32
+ "type_vocab_size": 2,
33
+ "vocab_size": 30000
34
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f69324d10cdc73964c32464f405ad0069c508e9f932626911a28d379bbc8d2e0
3
- size 267832560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55c21def950937d603e0c5209b8f4919e6ccd850fa82bf7502bb5b87db480922
3
+ size 46743912
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:370f97111278dc5bae72fbbf275aed53cb61ea6388e34a2a26d1d6ae9fb34780
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a89d981b6c77c4b865598567ba8078b5c04cddc89be6ec115b4c0d666b32c9c1
3
  size 4920