eskayML commited on
Commit
b83b328
·
verified ·
1 Parent(s): d1080a5

eskayML/electra_interview_new

Browse files
Files changed (5) hide show
  1. README.md +17 -17
  2. config.json +1 -1
  3. model.safetensors +1 -1
  4. tokenizer_config.json +1 -0
  5. training_args.bin +2 -2
README.md CHANGED
@@ -18,8 +18,8 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [mrm8488/electra-small-finetuned-squadv2](https://huggingface.co/mrm8488/electra-small-finetuned-squadv2) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 2.3959
22
- - Accuracy: 0.2675
23
 
24
  ## Model description
25
 
@@ -42,7 +42,7 @@ The following hyperparameters were used during training:
42
  - train_batch_size: 2
43
  - eval_batch_size: 2
44
  - seed: 42
45
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - num_epochs: 10
48
 
@@ -50,21 +50,21 @@ The following hyperparameters were used during training:
50
 
51
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
53
- | No log | 1.0 | 380 | 2.6074 | 0.2266 |
54
- | 2.7429 | 2.0 | 760 | 2.4872 | 0.2266 |
55
- | 2.5203 | 3.0 | 1140 | 2.4483 | 0.2266 |
56
- | 2.4479 | 4.0 | 1520 | 2.4349 | 0.2266 |
57
- | 2.4479 | 5.0 | 1900 | 2.4114 | 0.2306 |
58
- | 2.3919 | 6.0 | 2280 | 2.3933 | 0.2424 |
59
- | 2.2714 | 7.0 | 2660 | 2.3914 | 0.2530 |
60
- | 2.1536 | 8.0 | 3040 | 2.3968 | 0.2714 |
61
- | 2.1536 | 9.0 | 3420 | 2.3913 | 0.2648 |
62
- | 2.1058 | 10.0 | 3800 | 2.3959 | 0.2675 |
63
 
64
 
65
  ### Framework versions
66
 
67
- - Transformers 4.44.2
68
- - Pytorch 2.4.1+cu121
69
- - Datasets 3.0.1
70
- - Tokenizers 0.19.1
 
18
 
19
  This model is a fine-tuned version of [mrm8488/electra-small-finetuned-squadv2](https://huggingface.co/mrm8488/electra-small-finetuned-squadv2) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 2.2978
22
+ - Accuracy: 0.2716
23
 
24
  ## Model description
25
 
 
42
  - train_batch_size: 2
43
  - eval_batch_size: 2
44
  - seed: 42
45
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
  - num_epochs: 10
48
 
 
50
 
51
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
53
+ | No log | 1.0 | 463 | 2.6395 | 0.1983 |
54
+ | 2.7859 | 2.0 | 926 | 2.5260 | 0.1983 |
55
+ | 2.5604 | 3.0 | 1389 | 2.4446 | 0.2241 |
56
+ | 2.4612 | 4.0 | 1852 | 2.3737 | 0.3103 |
57
+ | 2.2886 | 5.0 | 2315 | 2.3307 | 0.3276 |
58
+ | 2.1381 | 6.0 | 2778 | 2.3076 | 0.3017 |
59
+ | 1.9905 | 7.0 | 3241 | 2.3089 | 0.2931 |
60
+ | 1.8363 | 8.0 | 3704 | 2.2939 | 0.2845 |
61
+ | 1.7738 | 9.0 | 4167 | 2.3060 | 0.2802 |
62
+ | 1.6807 | 10.0 | 4630 | 2.2978 | 0.2716 |
63
 
64
 
65
  ### Framework versions
66
 
67
+ - Transformers 4.47.1
68
+ - Pytorch 2.5.1+cu121
69
+ - Datasets 3.2.0
70
+ - Tokenizers 0.21.0
config.json CHANGED
@@ -68,7 +68,7 @@
68
  "summary_type": "first",
69
  "summary_use_proj": true,
70
  "torch_dtype": "float32",
71
- "transformers_version": "4.44.2",
72
  "type_vocab_size": 2,
73
  "use_cache": true,
74
  "vocab_size": 30522
 
68
  "summary_type": "first",
69
  "summary_use_proj": true,
70
  "torch_dtype": "float32",
71
+ "transformers_version": "4.47.1",
72
  "type_vocab_size": 2,
73
  "use_cache": true,
74
  "vocab_size": 30522
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:32a787c62a68c4edf191c25b81828f9bfa6b04143e2c06ea1ca0a61109ec1912
3
  size 54239712
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3894cfe4f3735a911a25d70b1c53da6bae76d75cbf5fd9839564037f9b0f8b89
3
  size 54239712
tokenizer_config.json CHANGED
@@ -45,6 +45,7 @@
45
  "cls_token": "[CLS]",
46
  "do_basic_tokenize": true,
47
  "do_lower_case": true,
 
48
  "mask_token": "[MASK]",
49
  "max_length": 512,
50
  "model_max_length": 512,
 
45
  "cls_token": "[CLS]",
46
  "do_basic_tokenize": true,
47
  "do_lower_case": true,
48
+ "extra_special_tokens": {},
49
  "mask_token": "[MASK]",
50
  "max_length": 512,
51
  "model_max_length": 512,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87d473bb239836d3c278ca4bc54501ef0abc689b9f6d2463e5a0867d8caa5240
3
- size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16905a4f2f0464a8209f41eeb68c1d827eaebd9b151a36821496df2a75c6de56
3
+ size 5304