Update config.json
#4
by
arda13olmez
- opened
- config.json +2 -2
config.json
CHANGED
@@ -84,7 +84,7 @@
|
|
84 |
"intermediate_size": 3072,
|
85 |
"num_attention_heads": 12,
|
86 |
"num_hidden_layers": 12,
|
87 |
-
"projection_dim": 768
|
88 |
},
|
89 |
"torch_dtype": "float32",
|
90 |
"transformers_version": null,
|
@@ -166,6 +166,6 @@
|
|
166 |
"num_attention_heads": 16,
|
167 |
"num_hidden_layers": 24,
|
168 |
"patch_size": 14,
|
169 |
-
"projection_dim": 768
|
170 |
}
|
171 |
}
|
|
|
84 |
"intermediate_size": 3072,
|
85 |
"num_attention_heads": 12,
|
86 |
"num_hidden_layers": 12,
|
87 |
+
"projection_dim": 768
|
88 |
},
|
89 |
"torch_dtype": "float32",
|
90 |
"transformers_version": null,
|
|
|
166 |
"num_attention_heads": 16,
|
167 |
"num_hidden_layers": 24,
|
168 |
"patch_size": 14,
|
169 |
+
"projection_dim": 768
|
170 |
}
|
171 |
}
|