DouglasPontes commited on
Commit
82cdd40
1 Parent(s): a72c9f5

Training in progress, step 32000

Browse files
README.md ADDED
@@ -0,0 +1,355 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: cardiffnlp/twitter-roberta-base-2019-90m
3
+ tags:
4
+ - generated_from_trainer
5
+ model-index:
6
+ - name: 2020-Q4-75p-filtered
7
+ results: []
8
+ ---
9
+
10
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
+ should probably proofread and complete it, then remove this comment. -->
12
+
13
+ # 2020-Q4-75p-filtered
14
+
15
+ This model is a fine-tuned version of [cardiffnlp/twitter-roberta-base-2019-90m](https://huggingface.co/cardiffnlp/twitter-roberta-base-2019-90m) on an unknown dataset.
16
+ It achieves the following results on the evaluation set:
17
+ - Loss: 3.0907
18
+
19
+ ## Model description
20
+
21
+ More information needed
22
+
23
+ ## Intended uses & limitations
24
+
25
+ More information needed
26
+
27
+ ## Training and evaluation data
28
+
29
+ More information needed
30
+
31
+ ## Training procedure
32
+
33
+ ### Training hyperparameters
34
+
35
+ The following hyperparameters were used during training:
36
+ - learning_rate: 4.1e-07
37
+ - train_batch_size: 16
38
+ - eval_batch_size: 16
39
+ - seed: 42
40
+ - optimizer: Adam with betas=(0.9,0.98) and epsilon=1e-08
41
+ - lr_scheduler_type: linear
42
+ - training_steps: 2400000
43
+
44
+ ### Training results
45
+
46
+ | Training Loss | Epoch | Step | Validation Loss |
47
+ |:-------------:|:-----:|:-------:|:---------------:|
48
+ | No log | 0.07 | 8000 | 3.5132 |
49
+ | 3.7013 | 0.13 | 16000 | 3.4076 |
50
+ | 3.7013 | 0.2 | 24000 | 3.3552 |
51
+ | 3.4836 | 0.27 | 32000 | 3.3266 |
52
+ | 3.4836 | 0.34 | 40000 | 3.3065 |
53
+ | 3.4297 | 0.4 | 48000 | 3.2915 |
54
+ | 3.4297 | 0.47 | 56000 | 3.2709 |
55
+ | 3.4017 | 0.54 | 64000 | 3.2550 |
56
+ | 3.4017 | 0.61 | 72000 | 3.2550 |
57
+ | 3.3841 | 0.67 | 80000 | 3.2372 |
58
+ | 3.3841 | 0.74 | 88000 | 3.2474 |
59
+ | 3.3749 | 0.81 | 96000 | 3.2446 |
60
+ | 3.3749 | 0.88 | 104000 | 3.2282 |
61
+ | 3.36 | 0.94 | 112000 | 3.2236 |
62
+ | 3.36 | 1.01 | 120000 | 3.2212 |
63
+ | 3.3556 | 1.08 | 128000 | 3.2011 |
64
+ | 3.3556 | 1.15 | 136000 | 3.2182 |
65
+ | 3.3516 | 1.21 | 144000 | 3.2046 |
66
+ | 3.3516 | 1.28 | 152000 | 3.2080 |
67
+ | 3.3485 | 1.35 | 160000 | 3.2197 |
68
+ | 3.3485 | 1.41 | 168000 | 3.1972 |
69
+ | 3.3429 | 1.48 | 176000 | 3.1990 |
70
+ | 3.3429 | 1.55 | 184000 | 3.1921 |
71
+ | 3.3383 | 1.62 | 192000 | 3.1891 |
72
+ | 3.3383 | 1.68 | 200000 | 3.1740 |
73
+ | 3.3352 | 1.75 | 208000 | 3.2062 |
74
+ | 3.3352 | 1.82 | 216000 | 3.1934 |
75
+ | 3.3376 | 1.89 | 224000 | 3.1893 |
76
+ | 3.3376 | 1.95 | 232000 | 3.1887 |
77
+ | 3.3272 | 2.02 | 240000 | 3.1821 |
78
+ | 3.3272 | 2.09 | 248000 | 3.1921 |
79
+ | 3.3329 | 2.16 | 256000 | 3.1812 |
80
+ | 3.3329 | 2.22 | 264000 | 3.1753 |
81
+ | 3.3241 | 2.29 | 272000 | 3.1790 |
82
+ | 3.3241 | 2.36 | 280000 | 3.1863 |
83
+ | 3.3271 | 2.43 | 288000 | 3.1850 |
84
+ | 3.3271 | 2.49 | 296000 | 3.1801 |
85
+ | 3.3223 | 2.56 | 304000 | 3.1687 |
86
+ | 3.3223 | 2.63 | 312000 | 3.1893 |
87
+ | 3.3211 | 2.69 | 320000 | 3.1691 |
88
+ | 3.3211 | 2.76 | 328000 | 3.1733 |
89
+ | 3.3226 | 2.83 | 336000 | 3.1659 |
90
+ | 3.3226 | 2.9 | 344000 | 3.1714 |
91
+ | 3.322 | 2.96 | 352000 | 3.1815 |
92
+ | 3.322 | 3.03 | 360000 | 3.1711 |
93
+ | 3.3094 | 3.1 | 368000 | 3.1669 |
94
+ | 3.3094 | 3.17 | 376000 | 3.1664 |
95
+ | 3.3172 | 3.23 | 384000 | 3.1541 |
96
+ | 3.3172 | 3.3 | 392000 | 3.1747 |
97
+ | 3.3134 | 3.37 | 400000 | 3.1648 |
98
+ | 3.3134 | 3.44 | 408000 | 3.1585 |
99
+ | 3.3194 | 3.5 | 416000 | 3.1651 |
100
+ | 3.3194 | 3.57 | 424000 | 3.1618 |
101
+ | 3.3199 | 3.64 | 432000 | 3.1566 |
102
+ | 3.3199 | 3.71 | 440000 | 3.1475 |
103
+ | 3.3185 | 3.77 | 448000 | 3.1504 |
104
+ | 3.3185 | 3.84 | 456000 | 3.1551 |
105
+ | 3.3072 | 3.91 | 464000 | 3.1636 |
106
+ | 3.3072 | 3.97 | 472000 | 3.1428 |
107
+ | 3.3103 | 4.04 | 480000 | 3.1537 |
108
+ | 3.3103 | 4.11 | 488000 | 3.1576 |
109
+ | 3.3106 | 4.18 | 496000 | 3.1442 |
110
+ | 3.3106 | 4.24 | 504000 | 3.1485 |
111
+ | 3.3091 | 4.31 | 512000 | 3.1377 |
112
+ | 3.3091 | 4.38 | 520000 | 3.1325 |
113
+ | 3.3051 | 4.45 | 528000 | 3.1570 |
114
+ | 3.3051 | 4.51 | 536000 | 3.1509 |
115
+ | 3.3034 | 4.58 | 544000 | 3.1481 |
116
+ | 3.3034 | 4.65 | 552000 | 3.1442 |
117
+ | 3.3076 | 4.72 | 560000 | 3.1451 |
118
+ | 3.3076 | 4.78 | 568000 | 3.1394 |
119
+ | 3.307 | 4.85 | 576000 | 3.1405 |
120
+ | 3.307 | 4.92 | 584000 | 3.1407 |
121
+ | 3.2994 | 4.99 | 592000 | 3.1483 |
122
+ | 3.2994 | 5.05 | 600000 | 3.1332 |
123
+ | 3.3017 | 5.12 | 608000 | 3.1447 |
124
+ | 3.3017 | 5.19 | 616000 | 3.1423 |
125
+ | 3.2931 | 5.25 | 624000 | 3.1427 |
126
+ | 3.2931 | 5.32 | 632000 | 3.1359 |
127
+ | 3.3042 | 5.39 | 640000 | 3.1399 |
128
+ | 3.3042 | 5.46 | 648000 | 3.1370 |
129
+ | 3.3006 | 5.52 | 656000 | 3.1319 |
130
+ | 3.3006 | 5.59 | 664000 | 3.1309 |
131
+ | 3.2973 | 5.66 | 672000 | 3.1408 |
132
+ | 3.2973 | 5.73 | 680000 | 3.1272 |
133
+ | 3.3023 | 5.79 | 688000 | 3.1403 |
134
+ | 3.3023 | 5.86 | 696000 | 3.1422 |
135
+ | 3.2968 | 5.93 | 704000 | 3.1204 |
136
+ | 3.2968 | 6.0 | 712000 | 3.1341 |
137
+ | 3.2982 | 6.06 | 720000 | 3.1381 |
138
+ | 3.2982 | 6.13 | 728000 | 3.1435 |
139
+ | 3.296 | 6.2 | 736000 | 3.1425 |
140
+ | 3.296 | 6.27 | 744000 | 3.1182 |
141
+ | 3.2949 | 6.33 | 752000 | 3.1336 |
142
+ | 3.2949 | 6.4 | 760000 | 3.1314 |
143
+ | 3.2925 | 6.47 | 768000 | 3.1347 |
144
+ | 3.2925 | 6.53 | 776000 | 3.1296 |
145
+ | 3.2981 | 6.6 | 784000 | 3.1308 |
146
+ | 3.2981 | 6.67 | 792000 | 3.1367 |
147
+ | 3.2904 | 6.74 | 800000 | 3.1176 |
148
+ | 3.2904 | 6.8 | 808000 | 3.1444 |
149
+ | 3.2936 | 6.87 | 816000 | 3.1213 |
150
+ | 3.2936 | 6.94 | 824000 | 3.1236 |
151
+ | 3.2856 | 7.01 | 832000 | 3.1276 |
152
+ | 3.2856 | 7.07 | 840000 | 3.1220 |
153
+ | 3.2956 | 7.14 | 848000 | 3.1211 |
154
+ | 3.2956 | 7.21 | 856000 | 3.1225 |
155
+ | 3.2872 | 7.28 | 864000 | 3.1220 |
156
+ | 3.2872 | 7.34 | 872000 | 3.1148 |
157
+ | 3.291 | 7.41 | 880000 | 3.1317 |
158
+ | 3.291 | 7.48 | 888000 | 3.1204 |
159
+ | 3.2923 | 7.55 | 896000 | 3.1148 |
160
+ | 3.2923 | 7.61 | 904000 | 3.1196 |
161
+ | 3.294 | 7.68 | 912000 | 3.1232 |
162
+ | 3.294 | 7.75 | 920000 | 3.1184 |
163
+ | 3.2838 | 7.81 | 928000 | 3.1157 |
164
+ | 3.2838 | 7.88 | 936000 | 3.1112 |
165
+ | 3.2848 | 7.95 | 944000 | 3.1097 |
166
+ | 3.2848 | 8.02 | 952000 | 3.1136 |
167
+ | 3.2797 | 8.08 | 960000 | 3.1108 |
168
+ | 3.2797 | 8.15 | 968000 | 3.1149 |
169
+ | 3.2852 | 8.22 | 976000 | 3.1050 |
170
+ | 3.2852 | 8.29 | 984000 | 3.0962 |
171
+ | 3.2833 | 8.35 | 992000 | 3.1139 |
172
+ | 3.2833 | 8.42 | 1000000 | 3.1099 |
173
+ | 3.2819 | 8.49 | 1008000 | 3.1126 |
174
+ | 3.2819 | 8.56 | 1016000 | 3.1167 |
175
+ | 3.2825 | 8.62 | 1024000 | 3.1057 |
176
+ | 3.2825 | 8.69 | 1032000 | 3.1059 |
177
+ | 3.2744 | 8.76 | 1040000 | 3.1075 |
178
+ | 3.2744 | 8.83 | 1048000 | 3.1068 |
179
+ | 3.2864 | 8.89 | 1056000 | 3.1098 |
180
+ | 3.2864 | 8.96 | 1064000 | 3.1151 |
181
+ | 3.2783 | 9.03 | 1072000 | 3.1022 |
182
+ | 3.2783 | 9.09 | 1080000 | 3.1196 |
183
+ | 3.2865 | 9.16 | 1088000 | 3.1177 |
184
+ | 3.2865 | 9.23 | 1096000 | 3.0985 |
185
+ | 3.2799 | 9.3 | 1104000 | 3.1073 |
186
+ | 3.2799 | 9.36 | 1112000 | 3.1019 |
187
+ | 3.2754 | 9.43 | 1120000 | 3.1003 |
188
+ | 3.2754 | 9.5 | 1128000 | 3.1015 |
189
+ | 3.2654 | 9.57 | 1136000 | 3.1047 |
190
+ | 3.2654 | 9.63 | 1144000 | 3.0931 |
191
+ | 3.2775 | 9.7 | 1152000 | 3.1130 |
192
+ | 3.2775 | 9.77 | 1160000 | 3.1094 |
193
+ | 3.2811 | 9.84 | 1168000 | 3.0964 |
194
+ | 3.2811 | 9.9 | 1176000 | 3.1069 |
195
+ | 3.2745 | 9.97 | 1184000 | 3.0995 |
196
+ | 3.2745 | 10.04 | 1192000 | 3.1167 |
197
+ | 3.2811 | 10.11 | 1200000 | 3.1119 |
198
+ | 3.2811 | 10.17 | 1208000 | 3.1063 |
199
+ | 3.2761 | 10.24 | 1216000 | 3.1041 |
200
+ | 3.2761 | 10.31 | 1224000 | 3.1044 |
201
+ | 3.2786 | 10.37 | 1232000 | 3.1051 |
202
+ | 3.2786 | 10.44 | 1240000 | 3.1068 |
203
+ | 3.2746 | 10.51 | 1248000 | 3.1057 |
204
+ | 3.2746 | 10.58 | 1256000 | 3.1009 |
205
+ | 3.2777 | 10.64 | 1264000 | 3.1103 |
206
+ | 3.2777 | 10.71 | 1272000 | 3.1073 |
207
+ | 3.268 | 10.78 | 1280000 | 3.1044 |
208
+ | 3.268 | 10.85 | 1288000 | 3.0951 |
209
+ | 3.266 | 10.91 | 1296000 | 3.1160 |
210
+ | 3.266 | 10.98 | 1304000 | 3.1156 |
211
+ | 3.2689 | 11.05 | 1312000 | 3.1033 |
212
+ | 3.2689 | 11.12 | 1320000 | 3.0985 |
213
+ | 3.2732 | 11.18 | 1328000 | 3.1103 |
214
+ | 3.2732 | 11.25 | 1336000 | 3.1059 |
215
+ | 3.2732 | 11.32 | 1344000 | 3.0810 |
216
+ | 3.2732 | 11.39 | 1352000 | 3.1074 |
217
+ | 3.2735 | 11.45 | 1360000 | 3.1070 |
218
+ | 3.2735 | 11.52 | 1368000 | 3.0996 |
219
+ | 3.2794 | 11.59 | 1376000 | 3.0987 |
220
+ | 3.2794 | 11.65 | 1384000 | 3.0890 |
221
+ | 3.2695 | 11.72 | 1392000 | 3.0998 |
222
+ | 3.2695 | 11.79 | 1400000 | 3.1010 |
223
+ | 3.2725 | 11.86 | 1408000 | 3.0921 |
224
+ | 3.2725 | 11.92 | 1416000 | 3.0921 |
225
+ | 3.2757 | 11.99 | 1424000 | 3.0949 |
226
+ | 3.2757 | 12.06 | 1432000 | 3.0957 |
227
+ | 3.2757 | 12.13 | 1440000 | 3.0918 |
228
+ | 3.2757 | 12.19 | 1448000 | 3.1131 |
229
+ | 3.2582 | 12.26 | 1456000 | 3.1006 |
230
+ | 3.2582 | 12.33 | 1464000 | 3.0912 |
231
+ | 3.2741 | 12.4 | 1472000 | 3.0938 |
232
+ | 3.2741 | 12.46 | 1480000 | 3.0977 |
233
+ | 3.2725 | 12.53 | 1488000 | 3.0973 |
234
+ | 3.2725 | 12.6 | 1496000 | 3.1070 |
235
+ | 3.2658 | 12.67 | 1504000 | 3.0992 |
236
+ | 3.2658 | 12.73 | 1512000 | 3.0927 |
237
+ | 3.2706 | 12.8 | 1520000 | 3.1007 |
238
+ | 3.2706 | 12.87 | 1528000 | 3.1003 |
239
+ | 3.2617 | 12.93 | 1536000 | 3.0978 |
240
+ | 3.2617 | 13.0 | 1544000 | 3.1059 |
241
+ | 3.2698 | 13.07 | 1552000 | 3.1061 |
242
+ | 3.2698 | 13.14 | 1560000 | 3.0986 |
243
+ | 3.2721 | 13.2 | 1568000 | 3.1078 |
244
+ | 3.2721 | 13.27 | 1576000 | 3.0985 |
245
+ | 3.2703 | 13.34 | 1584000 | 3.1060 |
246
+ | 3.2703 | 13.41 | 1592000 | 3.0889 |
247
+ | 3.2632 | 13.47 | 1600000 | 3.0970 |
248
+ | 3.2632 | 13.54 | 1608000 | 3.0893 |
249
+ | 3.275 | 13.61 | 1616000 | 3.1048 |
250
+ | 3.275 | 13.68 | 1624000 | 3.0975 |
251
+ | 3.2692 | 13.74 | 1632000 | 3.1019 |
252
+ | 3.2692 | 13.81 | 1640000 | 3.0796 |
253
+ | 3.2703 | 13.88 | 1648000 | 3.0986 |
254
+ | 3.2703 | 13.95 | 1656000 | 3.1036 |
255
+ | 3.2703 | 14.01 | 1664000 | 3.0973 |
256
+ | 3.2703 | 14.08 | 1672000 | 3.0910 |
257
+ | 3.2621 | 14.15 | 1680000 | 3.0832 |
258
+ | 3.2621 | 14.21 | 1688000 | 3.0910 |
259
+ | 3.2719 | 14.28 | 1696000 | 3.0927 |
260
+ | 3.2719 | 14.35 | 1704000 | 3.0935 |
261
+ | 3.2764 | 14.42 | 1712000 | 3.0914 |
262
+ | 3.2764 | 14.48 | 1720000 | 3.1064 |
263
+ | 3.265 | 14.55 | 1728000 | 3.0977 |
264
+ | 3.265 | 14.62 | 1736000 | 3.0933 |
265
+ | 3.261 | 14.69 | 1744000 | 3.0969 |
266
+ | 3.261 | 14.75 | 1752000 | 3.0911 |
267
+ | 3.2757 | 14.82 | 1760000 | 3.0764 |
268
+ | 3.2757 | 14.89 | 1768000 | 3.0865 |
269
+ | 3.276 | 14.96 | 1776000 | 3.0892 |
270
+ | 3.276 | 15.02 | 1784000 | 3.0981 |
271
+ | 3.2633 | 15.09 | 1792000 | 3.0920 |
272
+ | 3.2633 | 15.16 | 1800000 | 3.0744 |
273
+ | 3.2668 | 15.23 | 1808000 | 3.0933 |
274
+ | 3.2668 | 15.29 | 1816000 | 3.0907 |
275
+ | 3.2687 | 15.36 | 1824000 | 3.0862 |
276
+ | 3.2687 | 15.43 | 1832000 | 3.0910 |
277
+ | 3.2748 | 15.49 | 1840000 | 3.0950 |
278
+ | 3.2748 | 15.56 | 1848000 | 3.0823 |
279
+ | 3.2521 | 15.63 | 1856000 | 3.0944 |
280
+ | 3.2521 | 15.7 | 1864000 | 3.0819 |
281
+ | 3.2621 | 15.76 | 1872000 | 3.0942 |
282
+ | 3.2621 | 15.83 | 1880000 | 3.0998 |
283
+ | 3.2676 | 15.9 | 1888000 | 3.1061 |
284
+ | 3.2676 | 15.97 | 1896000 | 3.0957 |
285
+ | 3.2717 | 16.03 | 1904000 | 3.0878 |
286
+ | 3.2717 | 16.1 | 1912000 | 3.0802 |
287
+ | 3.2631 | 16.17 | 1920000 | 3.0800 |
288
+ | 3.2631 | 16.24 | 1928000 | 3.0903 |
289
+ | 3.2634 | 16.3 | 1936000 | 3.0924 |
290
+ | 3.2634 | 16.37 | 1944000 | 3.0843 |
291
+ | 3.2701 | 16.44 | 1952000 | 3.0869 |
292
+ | 3.2701 | 16.51 | 1960000 | 3.0776 |
293
+ | 3.2665 | 16.57 | 1968000 | 3.0928 |
294
+ | 3.2665 | 16.64 | 1976000 | 3.0835 |
295
+ | 3.2649 | 16.71 | 1984000 | 3.1013 |
296
+ | 3.2649 | 16.78 | 1992000 | 3.0895 |
297
+ | 3.2655 | 16.84 | 2000000 | 3.0814 |
298
+ | 3.2655 | 16.91 | 2008000 | 3.1003 |
299
+ | 3.2657 | 16.98 | 2016000 | 3.0924 |
300
+ | 3.2657 | 17.04 | 2024000 | 3.0946 |
301
+ | 3.2625 | 17.11 | 2032000 | 3.0956 |
302
+ | 3.2625 | 17.18 | 2040000 | 3.0961 |
303
+ | 3.2748 | 17.25 | 2048000 | 3.0851 |
304
+ | 3.2748 | 17.31 | 2056000 | 3.1006 |
305
+ | 3.2767 | 17.38 | 2064000 | 3.0888 |
306
+ | 3.2767 | 17.45 | 2072000 | 3.0800 |
307
+ | 3.2659 | 17.52 | 2080000 | 3.0859 |
308
+ | 3.2659 | 17.58 | 2088000 | 3.0929 |
309
+ | 3.2652 | 17.65 | 2096000 | 3.0882 |
310
+ | 3.2652 | 17.72 | 2104000 | 3.0930 |
311
+ | 3.2627 | 17.79 | 2112000 | 3.0789 |
312
+ | 3.2627 | 17.85 | 2120000 | 3.0959 |
313
+ | 3.2645 | 17.92 | 2128000 | 3.0989 |
314
+ | 3.2645 | 17.99 | 2136000 | 3.0852 |
315
+ | 3.2635 | 18.06 | 2144000 | 3.0948 |
316
+ | 3.2635 | 18.12 | 2152000 | 3.0890 |
317
+ | 3.2684 | 18.19 | 2160000 | 3.0906 |
318
+ | 3.2684 | 18.26 | 2168000 | 3.1040 |
319
+ | 3.2547 | 18.32 | 2176000 | 3.0923 |
320
+ | 3.2547 | 18.39 | 2184000 | 3.0820 |
321
+ | 3.2636 | 18.46 | 2192000 | 3.0937 |
322
+ | 3.2636 | 18.53 | 2200000 | 3.0951 |
323
+ | 3.2702 | 18.59 | 2208000 | 3.0887 |
324
+ | 3.2702 | 18.66 | 2216000 | 3.1017 |
325
+ | 3.257 | 18.73 | 2224000 | 3.0875 |
326
+ | 3.257 | 18.8 | 2232000 | 3.0903 |
327
+ | 3.2608 | 18.86 | 2240000 | 3.0945 |
328
+ | 3.2608 | 18.93 | 2248000 | 3.0850 |
329
+ | 3.2635 | 19.0 | 2256000 | 3.0877 |
330
+ | 3.2635 | 19.07 | 2264000 | 3.0873 |
331
+ | 3.2673 | 19.13 | 2272000 | 3.0925 |
332
+ | 3.2673 | 19.2 | 2280000 | 3.0947 |
333
+ | 3.2569 | 19.27 | 2288000 | 3.0901 |
334
+ | 3.2569 | 19.34 | 2296000 | 3.0889 |
335
+ | 3.2641 | 19.4 | 2304000 | 3.0916 |
336
+ | 3.2641 | 19.47 | 2312000 | 3.1022 |
337
+ | 3.2735 | 19.54 | 2320000 | 3.0927 |
338
+ | 3.2735 | 19.6 | 2328000 | 3.0938 |
339
+ | 3.2629 | 19.67 | 2336000 | 3.0892 |
340
+ | 3.2629 | 19.74 | 2344000 | 3.0883 |
341
+ | 3.2707 | 19.81 | 2352000 | 3.0935 |
342
+ | 3.2707 | 19.87 | 2360000 | 3.0909 |
343
+ | 3.2595 | 19.94 | 2368000 | 3.0933 |
344
+ | 3.2595 | 20.01 | 2376000 | 3.0822 |
345
+ | 3.2614 | 20.08 | 2384000 | 3.0932 |
346
+ | 3.2614 | 20.14 | 2392000 | 3.0980 |
347
+ | 3.2545 | 20.21 | 2400000 | 3.0907 |
348
+
349
+
350
+ ### Framework versions
351
+
352
+ - Transformers 4.35.0.dev0
353
+ - Pytorch 2.0.1+cu117
354
+ - Datasets 2.14.5
355
+ - Tokenizers 0.14.0
added_tokens.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "</s>": 2,
3
+ "<mask>": 50264,
4
+ "<pad>": 1,
5
+ "<s>": 0,
6
+ "<unk>": 3
7
+ }
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "cardiffnlp/twitter-roberta-base-2019-90m",
3
+ "architectures": [
4
+ "RobertaForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "layer_norm_eps": 1e-05,
17
+ "max_position_embeddings": 514,
18
+ "model_type": "roberta",
19
+ "num_attention_heads": 12,
20
+ "num_hidden_layers": 12,
21
+ "pad_token_id": 1,
22
+ "position_embedding_type": "absolute",
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.35.0.dev0",
25
+ "type_vocab_size": 1,
26
+ "use_cache": true,
27
+ "vocab_size": 50265
28
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1274cf0c57a9557352d4303c9552e48c4d262e71b7ae5d2c3e40ed285614918a
3
+ size 498859189
special_tokens_map.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": "<mask>",
6
+ "pad_token": "<pad>",
7
+ "sep_token": "</s>",
8
+ "unk_token": "<unk>"
9
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
+ "additional_special_tokens": [],
46
+ "bos_token": "<s>",
47
+ "clean_up_tokenization_spaces": true,
48
+ "cls_token": "<s>",
49
+ "eos_token": "</s>",
50
+ "errors": "replace",
51
+ "mask_token": "<mask>",
52
+ "max_length": 512,
53
+ "model_max_length": 512,
54
+ "pad_token": "<pad>",
55
+ "sep_token": "</s>",
56
+ "stride": 0,
57
+ "tokenizer_class": "RobertaTokenizer",
58
+ "trim_offsets": true,
59
+ "truncation_side": "right",
60
+ "truncation_strategy": "longest_first",
61
+ "unk_token": "<unk>"
62
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5403c2d785b63d08ac67b0f7051dfa6f902095b3e5dd128051996f54d35c972
3
+ size 4219
vocab.json ADDED
The diff for this file is too large to render. See raw diff