casperhansen commited on
Commit
32eeeb5
1 Parent(s): afedc47

Hotfix for not saving correctly (#762)

Browse files
src/axolotl/monkeypatch/fused_modules.py DELETED
File without changes
src/axolotl/monkeypatch/llama_attn_hijack_flash.py CHANGED
@@ -152,6 +152,7 @@ class FusedAttention(LlamaAttention):
152
  new_attn.q_proj.weight.data = q_proj
153
  new_attn.k_proj.weight.data = k_proj
154
  new_attn.v_proj.weight.data = v_proj
 
155
 
156
  set_module_name(model, name, new_attn)
157
 
 
152
  new_attn.q_proj.weight.data = q_proj
153
  new_attn.k_proj.weight.data = k_proj
154
  new_attn.v_proj.weight.data = v_proj
155
+ new_attn.o_proj.weight.data = self.o_proj.weight.data
156
 
157
  set_module_name(model, name, new_attn)
158