ZinengTang commited on
Commit
16f7f07
1 Parent(s): aa30e6d

Upload model

Browse files
Files changed (2) hide show
  1. config.json +14 -18
  2. pytorch_model.bin +2 -2
config.json CHANGED
@@ -1,25 +1,21 @@
1
  {
2
  "architectures": [
3
- "Blip2QFormerModel"
4
  ],
5
- "num_query_tokens": 32,
6
- "attention_probs_dropout_prob": 0.1,
7
- "classifier_dropout": null,
8
- "cross_attention_frequency": 2,
9
- "encoder_hidden_size": 1408,
10
- "hidden_act": "gelu",
11
- "hidden_dropout_prob": 0.1,
12
- "hidden_size": 768,
13
  "initializer_range": 0.02,
14
- "intermediate_size": 3072,
15
- "layer_norm_eps": 1e-12,
16
- "max_position_embeddings": 512,
17
- "model_type": "blip_2_qformer",
18
- "num_attention_heads": 12,
19
- "num_hidden_layers": 12,
20
- "pad_token_id": 0,
21
- "position_embedding_type": "absolute",
22
  "torch_dtype": "float32",
23
  "transformers_version": "4.34.1",
24
- "vocab_size": 30522
 
 
 
25
  }
 
1
  {
2
  "architectures": [
3
+ "ImageQFormerModel"
4
  ],
5
+ "initializer_factor": 1.0,
 
 
 
 
 
 
 
6
  "initializer_range": 0.02,
7
+ "model_type": "blip-2",
8
+ "num_query_tokens": 32,
9
+ "qformer_config": {
10
+ "model_type": "blip_2_qformer"
11
+ },
12
+ "text_config": {
13
+ "model_type": "opt"
14
+ },
15
  "torch_dtype": "float32",
16
  "transformers_version": "4.34.1",
17
+ "use_decoder_only_language_model": true,
18
+ "vision_config": {
19
+ "model_type": "blip_2_vision_model"
20
+ }
21
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:96d286e0cda62601dfa1ff37ec4a73951af8c49644da3174e1b3b4146b8a7a86
3
- size 420732489
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e01c57813d103c4430468dca693984c0f547e3a7fed28ea17208997e6c5d31b0
3
+ size 4364700517