quim-motger commited on
Commit
f36f380
1 Parent(s): 3a93c4f

Upload 12 files

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "roberta-large",
3
  "architectures": [
4
  "RobertaForTokenClassification"
5
  ],
@@ -30,7 +30,7 @@
30
  "pad_token_id": 1,
31
  "position_embedding_type": "absolute",
32
  "torch_dtype": "float32",
33
- "transformers_version": "4.30.2",
34
  "type_vocab_size": 1,
35
  "use_cache": true,
36
  "vocab_size": 50265
 
1
  {
2
+ "_name_or_path": "data/further_pretraining/roberta-large/checkpoint-4678",
3
  "architectures": [
4
  "RobertaForTokenClassification"
5
  ],
 
30
  "pad_token_id": 1,
31
  "position_embedding_type": "absolute",
32
  "torch_dtype": "float32",
33
+ "transformers_version": "4.39.1",
34
  "type_vocab_size": 1,
35
  "use_cache": true,
36
  "vocab_size": 50265
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c819517d251abbd2ac4c61ff6ce7f58aa7de715f4d4f5af130f5ed066a713dec
3
+ size 1417300884
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8b3afdcf861ae2fb7615af0a9af193b0ed91811b51bf08df305bba7e79e2fb0
3
- size 2834735429
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cec5a9c7f1e183f8603f47c0daa973e2b992af8f0338375f08f1b12c43a2d087
3
+ size 2834832935
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:535668448099908a7c1073e58ae7c2fe30d9dd5b4d8377f0ad0f321344df25ba
3
- size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c42a88072d467bd76274de862e37ee53c3e1d7f33ad54c8b28c097a74bbcc46b
3
+ size 14244
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a81b9e64a6d36ab9d27ca33efaa732434676daf622cfdf89ee1179762e733965
3
- size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af8e36d67e7350aa7ba27dfd844d165ee5ffe669ec863b55108a58646f643dd9
3
+ size 1064
tokenizer.json CHANGED
@@ -14,7 +14,7 @@
14
  "single_word": false,
15
  "lstrip": false,
16
  "rstrip": false,
17
- "normalized": false,
18
  "special": true
19
  },
20
  {
@@ -23,7 +23,7 @@
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
26
- "normalized": false,
27
  "special": true
28
  },
29
  {
@@ -32,7 +32,7 @@
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
35
- "normalized": false,
36
  "special": true
37
  },
38
  {
@@ -41,7 +41,7 @@
41
  "single_word": false,
42
  "lstrip": false,
43
  "rstrip": false,
44
- "normalized": false,
45
  "special": true
46
  },
47
  {
 
14
  "single_word": false,
15
  "lstrip": false,
16
  "rstrip": false,
17
+ "normalized": true,
18
  "special": true
19
  },
20
  {
 
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
26
+ "normalized": true,
27
  "special": true
28
  },
29
  {
 
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
35
+ "normalized": true,
36
  "special": true
37
  },
38
  {
 
41
  "single_word": false,
42
  "lstrip": false,
43
  "rstrip": false,
44
+ "normalized": true,
45
  "special": true
46
  },
47
  {
tokenizer_config.json CHANGED
@@ -1,5 +1,47 @@
1
  {
2
  "add_prefix_space": true,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "bos_token": "<s>",
4
  "clean_up_tokenization_spaces": true,
5
  "cls_token": "<s>",
 
1
  {
2
  "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
  "bos_token": "<s>",
46
  "clean_up_tokenization_spaces": true,
47
  "cls_token": "<s>",
trainer_state.json CHANGED
@@ -1,64 +1,45 @@
1
  {
2
- "best_metric": 0.014085530303418636,
3
- "best_model_checkpoint": "data/train-test///model/checkpoint-3351",
4
- "epoch": 1.0,
5
- "global_step": 3351,
 
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
- "epoch": 0.15,
12
- "learning_rate": 1.850790808713817e-05,
13
- "loss": 0.0395,
14
- "step": 500
15
- },
16
- {
17
- "epoch": 0.3,
18
- "learning_rate": 1.7015816174276335e-05,
19
- "loss": 0.0125,
20
- "step": 1000
21
- },
22
- {
23
- "epoch": 0.45,
24
- "learning_rate": 1.5523724261414504e-05,
25
- "loss": 0.009,
26
- "step": 1500
27
- },
28
- {
29
- "epoch": 0.6,
30
- "learning_rate": 1.4031632348552673e-05,
31
- "loss": 0.0056,
32
- "step": 2000
33
- },
34
- {
35
- "epoch": 0.75,
36
- "learning_rate": 1.253954043569084e-05,
37
- "loss": 0.0056,
38
- "step": 2500
39
- },
40
- {
41
- "epoch": 0.9,
42
- "learning_rate": 1.1047448522829008e-05,
43
- "loss": 0.0197,
44
- "step": 3000
45
  },
46
  {
47
- "epoch": 1.0,
48
  "eval_accuracy": null,
49
- "eval_f1": 0.9771134983652499,
50
- "eval_loss": 0.014085530303418636,
51
- "eval_precision": 0.9840075258701787,
52
- "eval_recall": 0.9703153988868275,
53
- "eval_runtime": 1.14,
54
- "eval_samples_per_second": 850.884,
55
- "eval_steps_per_second": 107.018,
56
- "step": 3351
57
  }
58
  ],
59
- "max_steps": 6702,
 
 
60
  "num_train_epochs": 2,
61
- "total_flos": 2207357174625474.0,
 
 
62
  "trial_name": null,
63
  "trial_params": null
64
  }
 
1
  {
2
+ "best_metric": 0.937442502299908,
3
+ "best_model_checkpoint": "data/train-test/roberta-large-output//model/checkpoint-232",
4
+ "epoch": 2.0,
5
+ "eval_steps": 500,
6
+ "global_step": 232,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 1.0,
13
+ "eval_accuracy": null,
14
+ "eval_f1": 0.902638762511374,
15
+ "eval_loss": 0.05000825226306915,
16
+ "eval_precision": 0.8928892889288929,
17
+ "eval_recall": 0.9126034958601656,
18
+ "eval_runtime": 4.4114,
19
+ "eval_samples_per_second": 219.883,
20
+ "eval_steps_per_second": 7.027,
21
+ "step": 116
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
  },
23
  {
24
+ "epoch": 2.0,
25
  "eval_accuracy": null,
26
+ "eval_f1": 0.9276285844333181,
27
+ "eval_loss": 0.03446832671761513,
28
+ "eval_precision": 0.918018018018018,
29
+ "eval_recall": 0.937442502299908,
30
+ "eval_runtime": 4.4048,
31
+ "eval_samples_per_second": 220.215,
32
+ "eval_steps_per_second": 7.038,
33
+ "step": 232
34
  }
35
  ],
36
+ "logging_steps": 500,
37
+ "max_steps": 232,
38
+ "num_input_tokens_seen": 0,
39
  "num_train_epochs": 2,
40
+ "save_steps": 500,
41
+ "total_flos": 853475550077862.0,
42
+ "train_batch_size": 32,
43
  "trial_name": null,
44
  "trial_params": null
45
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bb8bb5ea0455a2f363fb152a305bef728199f81ce224543eb10a042f60f7272b
3
- size 3963
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b250203d90fac6c5a543f3f4385532a2720339f6f459179f95dd9d53b9e68027
3
+ size 4984