crumb commited on
Commit
1ee6f60
1 Parent(s): e27b807
Files changed (1) hide show
  1. modeling_t2.py +2 -0
modeling_t2.py CHANGED
@@ -149,6 +149,8 @@ class TransformerAttention(nn.Module):
149
  super().__init__()
150
  self.config = config
151
  self.stage = stage
 
 
152
  self.head_dim = config.hidden_size // config.num_attention_heads
153
  assert (
154
  self.head_dim * config.num_attention_heads == config.hidden_size
 
149
  super().__init__()
150
  self.config = config
151
  self.stage = stage
152
+ self.max_position_embeddings = config.max_position_embeddings
153
+ self.rope_theta = config.rope_theta
154
  self.head_dim = config.hidden_size // config.num_attention_heads
155
  assert (
156
  self.head_dim * config.num_attention_heads == config.hidden_size