revert previous change and build ax images w docker on gpu (#371) 918f1b0 unverified winglian commited on Aug 13, 2023
attempt to run non-base docker builds on regular cpu hosts (#369) c3fde36 unverified winglian commited on Aug 12, 2023
Attention mask and position id fixes for packing (#285) 2bb0b78 unverified winglian commited on Aug 12, 2023
Add wandb_entity to wandb options, update example configs, update README (#361) 7019509 unverified Morgan McGuire Morgan McGuire winglian commited on Aug 12, 2023
Fix(model loading): Warn when model revision is passed to gptq (#364) 96bd6ae unverified Nanobit commited on Aug 12, 2023
Fix(message): Improve error message for bad format (#365) e37d935 unverified Nanobit commited on Aug 12, 2023
Merge pull request #355 from tmm1/bitsandbytes-fixes 35c8b90 unverified tmm1 commited on Aug 11, 2023
Merge pull request #350 from tmm1/group-len-false-examples f5c11f8 unverified tmm1 commited on Aug 9, 2023
ensure enable_input_require_grads is called on model before getting the peft model (#345) 176b888 unverified winglian commited on Aug 6, 2023
experimental llama 2 chat support (#296) 3392270 unverified Jan Philipp Harries Jan Philipp Harries commited on Aug 6, 2023
Update XFormers Attention Monkeypatch to handle Llama-2 70B (GQA) (#339) 10405b9 unverified ssmi153 commited on Aug 6, 2023
Added Orca Mini prompt strategy (#263) c93655c unverified Jan Philipp Harries Jan Philipp Harries commited on Aug 5, 2023
optimize the iteration when tokenizeing large datasets (#332) fe28543 unverified winglian commited on Aug 4, 2023
scope flash-attn+qlora fix correctly, scope to llama, add comment 78b9efb tmm1 commited on Aug 3, 2023
ensure flash-attn fixes happen in both adapter/lora modes, and use torch_dtype 248bf90 tmm1 commited on Aug 2, 2023
add peft install back since it doesn't get installed by setup.py (#331) db2a358 unverified winglian commited on Jul 31, 2023