eskayML commited on
Commit
ce1aaab
·
verified ·
1 Parent(s): 74772fc

eskayML/interview_electra

Browse files
Files changed (4) hide show
  1. README.md +13 -18
  2. config.json +7 -41
  3. model.safetensors +2 -2
  4. training_args.bin +2 -2
README.md CHANGED
@@ -18,8 +18,8 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [mrm8488/electra-small-finetuned-squadv2](https://huggingface.co/mrm8488/electra-small-finetuned-squadv2) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 2.4751
22
- - Accuracy: 0.2305
23
 
24
  ## Model description
25
 
@@ -42,29 +42,24 @@ The following hyperparameters were used during training:
42
  - train_batch_size: 2
43
  - eval_batch_size: 2
44
  - seed: 42
45
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
- - num_epochs: 10
48
 
49
  ### Training results
50
 
51
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
53
- | No log | 1.0 | 152 | 2.8162 | 0.2272 |
54
- | No log | 2.0 | 304 | 2.6814 | 0.2263 |
55
- | No log | 3.0 | 456 | 2.5987 | 0.2263 |
56
- | 2.759 | 4.0 | 608 | 2.5461 | 0.2263 |
57
- | 2.759 | 5.0 | 760 | 2.5203 | 0.2272 |
58
- | 2.759 | 6.0 | 912 | 2.4958 | 0.2272 |
59
- | 2.5023 | 7.0 | 1064 | 2.4920 | 0.2305 |
60
- | 2.5023 | 8.0 | 1216 | 2.4850 | 0.2305 |
61
- | 2.5023 | 9.0 | 1368 | 2.4786 | 0.2296 |
62
- | 2.372 | 10.0 | 1520 | 2.4751 | 0.2305 |
63
 
64
 
65
  ### Framework versions
66
 
67
- - Transformers 4.44.2
68
- - Pytorch 2.4.1+cu121
69
- - Datasets 3.0.1
70
- - Tokenizers 0.19.1
 
18
 
19
  This model is a fine-tuned version of [mrm8488/electra-small-finetuned-squadv2](https://huggingface.co/mrm8488/electra-small-finetuned-squadv2) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.8333
22
+ - Accuracy: 0.6316
23
 
24
  ## Model description
25
 
 
42
  - train_batch_size: 2
43
  - eval_batch_size: 2
44
  - seed: 42
45
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
+ - num_epochs: 5
48
 
49
  ### Training results
50
 
51
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
53
+ | No log | 1.0 | 95 | 1.0014 | 0.5526 |
54
+ | No log | 2.0 | 190 | 0.9542 | 0.5526 |
55
+ | No log | 3.0 | 285 | 0.9142 | 0.5526 |
56
+ | No log | 4.0 | 380 | 0.8572 | 0.6316 |
57
+ | No log | 5.0 | 475 | 0.8333 | 0.6316 |
 
 
 
 
 
58
 
59
 
60
  ### Framework versions
61
 
62
+ - Transformers 4.46.2
63
+ - Pytorch 2.5.1+cu121
64
+ - Datasets 3.1.0
65
+ - Tokenizers 0.20.3
config.json CHANGED
@@ -10,50 +10,16 @@
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 256,
12
  "id2label": {
13
- "0": "ALGORITHM",
14
- "1": "COMANAGEMENT",
15
- "2": "COMPETING PRIORITIES",
16
- "3": "COST",
17
- "4": "COVID",
18
- "5": "EHR",
19
- "6": "GUIDELINES",
20
- "7": "HFREF COMFORT MANAGING",
21
- "8": "INFO NEEDS",
22
- "9": "INTERNAL & EXTERNAL SUPPORT BAR",
23
- "10": "JUICY QUOTE",
24
- "11": "MAGIC WAND",
25
- "12": "MED MANAGEMENT",
26
- "13": "OTHER",
27
- "14": "PATIENT DIGITAL TOOLS",
28
- "15": "PATIENTS",
29
- "16": "PROVIDER EDUCATION",
30
- "17": "ROLE OF OTHER STAFF",
31
- "18": "TIME CONSTRAINTS",
32
- "19": "WORKFLOW"
33
  },
34
  "initializer_range": 0.02,
35
  "intermediate_size": 1024,
36
  "label2id": {
37
- "ALGORITHM": 0,
38
- "COMANAGEMENT": 1,
39
- "COMPETING PRIORITIES": 2,
40
- "COST": 3,
41
- "COVID": 4,
42
- "EHR": 5,
43
- "GUIDELINES": 6,
44
- "HFREF COMFORT MANAGING": 7,
45
- "INFO NEEDS": 8,
46
- "INTERNAL & EXTERNAL SUPPORT BAR": 9,
47
- "JUICY QUOTE": 10,
48
- "MAGIC WAND": 11,
49
- "MED MANAGEMENT": 12,
50
- "OTHER": 13,
51
- "PATIENT DIGITAL TOOLS": 14,
52
- "PATIENTS": 15,
53
- "PROVIDER EDUCATION": 16,
54
- "ROLE OF OTHER STAFF": 17,
55
- "TIME CONSTRAINTS": 18,
56
- "WORKFLOW": 19
57
  },
58
  "layer_norm_eps": 1e-12,
59
  "max_position_embeddings": 512,
@@ -68,7 +34,7 @@
68
  "summary_type": "first",
69
  "summary_use_proj": true,
70
  "torch_dtype": "float32",
71
- "transformers_version": "4.44.2",
72
  "type_vocab_size": 2,
73
  "use_cache": true,
74
  "vocab_size": 30522
 
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 256,
12
  "id2label": {
13
+ "0": "multi_level_org_char",
14
+ "1": "multi_level_org_perspect",
15
+ "2": "impl_sust_infra"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  },
17
  "initializer_range": 0.02,
18
  "intermediate_size": 1024,
19
  "label2id": {
20
+ "impl_sust_infra": 2,
21
+ "multi_level_org_char": 0,
22
+ "multi_level_org_perspect": 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  },
24
  "layer_norm_eps": 1e-12,
25
  "max_position_embeddings": 512,
 
34
  "summary_type": "first",
35
  "summary_use_proj": true,
36
  "torch_dtype": "float32",
37
+ "transformers_version": "4.46.2",
38
  "type_vocab_size": 2,
39
  "use_cache": true,
40
  "vocab_size": 30522
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:309612d49ad131b9df62294bef9d1d48285a346e9121559ed563704470432338
3
- size 54239712
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd26f6c33f8d77ba64f4be863e580323fb8e5d14d6ad07950c3e89acb0f0b6b6
3
+ size 54222228
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:59c7749d1812864cdb82fd1553bf513392820fa93c6e6061a9a65e2909020c8e
3
- size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd78eec26b21fb161eccebcfe70951605ce8b107fdcc84f5c0b455589be01483
3
+ size 5240