Files changed (1) hide show
  1. config.json +2 -2
config.json CHANGED
@@ -84,7 +84,7 @@
84
  "intermediate_size": 3072,
85
  "num_attention_heads": 12,
86
  "num_hidden_layers": 12,
87
- "projection_dim": 768,
88
  },
89
  "torch_dtype": "float32",
90
  "transformers_version": null,
@@ -166,6 +166,6 @@
166
  "num_attention_heads": 16,
167
  "num_hidden_layers": 24,
168
  "patch_size": 14,
169
- "projection_dim": 768,
170
  }
171
  }
 
84
  "intermediate_size": 3072,
85
  "num_attention_heads": 12,
86
  "num_hidden_layers": 12,
87
+ "projection_dim": 768
88
  },
89
  "torch_dtype": "float32",
90
  "transformers_version": null,
 
166
  "num_attention_heads": 16,
167
  "num_hidden_layers": 24,
168
  "patch_size": 14,
169
+ "projection_dim": 768
170
  }
171
  }