Update to fix error when using with transformers
#14
by
dome272
- opened
- config.json +1 -0
config.json
CHANGED
@@ -30,6 +30,7 @@
|
|
30 |
"forced_eos_token_id": null,
|
31 |
"hidden_act": "gelu",
|
32 |
"hidden_size": 1280,
|
|
|
33 |
"id2label": {
|
34 |
"0": "LABEL_0",
|
35 |
"1": "LABEL_1"
|
|
|
30 |
"forced_eos_token_id": null,
|
31 |
"hidden_act": "gelu",
|
32 |
"hidden_size": 1280,
|
33 |
+
"projection_dim": 1280,
|
34 |
"id2label": {
|
35 |
"0": "LABEL_0",
|
36 |
"1": "LABEL_1"
|