Update configuration_intern_vit.py
Browse files
configuration_intern_vit.py
CHANGED
@@ -73,6 +73,7 @@ class InternVisionConfig(PretrainedConfig):
|
|
73 |
num_hidden_layers=48,
|
74 |
use_flash_attn=True,
|
75 |
hidden_act='gelu',
|
|
|
76 |
layer_norm_eps=1e-6,
|
77 |
dropout=0.0,
|
78 |
drop_path_rate=0.0,
|
@@ -97,6 +98,7 @@ class InternVisionConfig(PretrainedConfig):
|
|
97 |
self.attention_dropout = attention_dropout
|
98 |
self.layer_norm_eps = layer_norm_eps
|
99 |
self.hidden_act = hidden_act
|
|
|
100 |
self.qkv_bias = qkv_bias
|
101 |
self.qk_normalization = qk_normalization
|
102 |
self.use_flash_attn = use_flash_attn
|
|
|
73 |
num_hidden_layers=48,
|
74 |
use_flash_attn=True,
|
75 |
hidden_act='gelu',
|
76 |
+
norm_type='rms_norm',
|
77 |
layer_norm_eps=1e-6,
|
78 |
dropout=0.0,
|
79 |
drop_path_rate=0.0,
|
|
|
98 |
self.attention_dropout = attention_dropout
|
99 |
self.layer_norm_eps = layer_norm_eps
|
100 |
self.hidden_act = hidden_act
|
101 |
+
self.norm_type = norm_type
|
102 |
self.qkv_bias = qkv_bias
|
103 |
self.qk_normalization = qk_normalization
|
104 |
self.use_flash_attn = use_flash_attn
|