murthyrudra commited on
Commit
8d98128
1 Parent(s): 7dafff2

End of training

Browse files
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
GOAT ADDED
@@ -0,0 +1 @@
 
 
1
+ Subproject commit 7dafff2a6315310e8dc30b73951fbc0a05fdd76d
config.json ADDED
@@ -0,0 +1,78 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "../HiNER/cfilt_original/cfilt_original-xlm-roberta-large_ner_16_3e-05_10_1/",
3
+ "architectures": [
4
+ "XLMRobertaForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 1024,
13
+ "id2label": {
14
+ "0": "B-FESTIVAL",
15
+ "1": "B-GAME",
16
+ "2": "B-LANGUAGE",
17
+ "3": "B-LITERATURE",
18
+ "4": "B-LOCATION",
19
+ "5": "B-MISC",
20
+ "6": "B-NUMEX",
21
+ "7": "B-ORGANIZATION",
22
+ "8": "B-PERSON",
23
+ "9": "B-RELIGION",
24
+ "10": "B-TIMEX",
25
+ "11": "I-FESTIVAL",
26
+ "12": "I-GAME",
27
+ "13": "I-LANGUAGE",
28
+ "14": "I-LITERATURE",
29
+ "15": "I-LOCATION",
30
+ "16": "I-MISC",
31
+ "17": "I-NUMEX",
32
+ "18": "I-ORGANIZATION",
33
+ "19": "I-PERSON",
34
+ "20": "I-RELIGION",
35
+ "21": "I-TIMEX",
36
+ "22": "O"
37
+ },
38
+ "initializer_range": 0.02,
39
+ "intermediate_size": 4096,
40
+ "label2id": {
41
+ "B-FESTIVAL": 0,
42
+ "B-GAME": 1,
43
+ "B-LANGUAGE": 2,
44
+ "B-LITERATURE": 3,
45
+ "B-LOCATION": 4,
46
+ "B-MISC": 5,
47
+ "B-NUMEX": 6,
48
+ "B-ORGANIZATION": 7,
49
+ "B-PERSON": 8,
50
+ "B-RELIGION": 9,
51
+ "B-TIMEX": 10,
52
+ "I-FESTIVAL": 11,
53
+ "I-GAME": 12,
54
+ "I-LANGUAGE": 13,
55
+ "I-LITERATURE": 14,
56
+ "I-LOCATION": 15,
57
+ "I-MISC": 16,
58
+ "I-NUMEX": 17,
59
+ "I-ORGANIZATION": 18,
60
+ "I-PERSON": 19,
61
+ "I-RELIGION": 20,
62
+ "I-TIMEX": 21,
63
+ "O": 22
64
+ },
65
+ "layer_norm_eps": 1e-05,
66
+ "max_position_embeddings": 514,
67
+ "model_type": "xlm-roberta",
68
+ "num_attention_heads": 16,
69
+ "num_hidden_layers": 24,
70
+ "output_past": true,
71
+ "pad_token_id": 1,
72
+ "position_embedding_type": "absolute",
73
+ "torch_dtype": "float32",
74
+ "transformers_version": "4.14.0",
75
+ "type_vocab_size": 1,
76
+ "use_cache": true,
77
+ "vocab_size": 250002
78
+ }
dev_predictions.txt ADDED
The diff for this file is too large to render. See raw diff
 
dev_results.txt ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ test_loss = 0.09286307543516159
2
+ test_precision = 0.894966767560146
3
+ test_recall = 0.8859269784394885
4
+ test_f1 = 0.8904239300849103
5
+ test_runtime = 160.3832
6
+ test_samples_per_second = 68.056
7
+ test_steps_per_second = 8.511
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e73accfabb54e8675e7e5422cf735faace4ff48e4b13b274f75e0674c3a14e25
3
+ size 2235621041
runs/May01_03-39-27_cccxc557/events.out.tfevents.1651391671.cccxc557.2725506.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1021e213e29038d11ea30daf59d0a79d09186c7046f04c11cedce2ff34b9b968
3
+ size 389
test_predictions.txt ADDED
The diff for this file is too large to render. See raw diff
 
test_results.txt ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ test_loss = 0.08977837860584259
2
+ test_precision = 0.8968858782575971
3
+ test_recall = 0.8871207891308394
4
+ test_f1 = 0.8919766081871345
5
+ test_runtime = 315.5399
6
+ test_samples_per_second = 69.183
7
+ test_steps_per_second = 8.649
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e45fe6d9e9dfd02e8d2b0808300fe72e0ec2ba85806570d64852b665447c00d
3
+ size 3055