;)
Browse files- modeling_t2.py +2 -0
modeling_t2.py
CHANGED
@@ -149,6 +149,8 @@ class TransformerAttention(nn.Module):
|
|
149 |
super().__init__()
|
150 |
self.config = config
|
151 |
self.stage = stage
|
|
|
|
|
152 |
self.head_dim = config.hidden_size // config.num_attention_heads
|
153 |
assert (
|
154 |
self.head_dim * config.num_attention_heads == config.hidden_size
|
|
|
149 |
super().__init__()
|
150 |
self.config = config
|
151 |
self.stage = stage
|
152 |
+
self.max_position_embeddings = config.max_position_embeddings
|
153 |
+
self.rope_theta = config.rope_theta
|
154 |
self.head_dim = config.hidden_size // config.num_attention_heads
|
155 |
assert (
|
156 |
self.head_dim * config.num_attention_heads == config.hidden_size
|