Add training callback to send predictions to WandB table (#521) 5b67ea9 unverified Glavin001 commited on Sep 13, 2023
Add support for GPTQ using native transformers/peft (#468) 3355706 unverified winglian commited on Sep 5, 2023
fsdp requires params be the same type too (#493) 98bf76e unverified winglian commited on Aug 28, 2023
Fix(tokenizer): Make sure to add pad for CodeLlamaTokenizer (#489) 4c37bd0 unverified Nanobit commited on Aug 28, 2023
fix: finetune model inference needs the dtype fix to work with flash-attn f311df9 unverified Maxime commited on Aug 26, 2023
Fix(tokenizer): Fix condition to add pad token (#477) 71bd062 unverified Nanobit commited on Aug 25, 2023
recast loralayer, norm, lmhead + embed token weights per original qlora (#393) 96deb6b unverified winglian commited on Aug 21, 2023
don't pass rope_scaling kwarg if it's None (#383) 919246f unverified winglian commited on Aug 13, 2023
try to detect accelerate and only use device_map=None in that case (#373) 094fc2c unverified tmm1 commited on Aug 13, 2023
Attention mask and position id fixes for packing (#285) 2bb0b78 unverified winglian commited on Aug 12, 2023
ensure enable_input_require_grads is called on model before getting the peft model (#345) 176b888 unverified winglian commited on Aug 6, 2023
scope flash-attn+qlora fix correctly, scope to llama, add comment 78b9efb tmm1 commited on Aug 3, 2023
ensure flash-attn fixes happen in both adapter/lora modes, and use torch_dtype 248bf90 tmm1 commited on Aug 2, 2023
add peft install back since it doesn't get installed by setup.py (#331) db2a358 unverified winglian commited on Jul 31, 2023
don't use llama if trust_remote_code is set since that needs to use AutoModel path 66afb76 winglian commited on Jul 8, 2023