Mistral: Sliding Window Attention with Flash Attention and Sample Packing (#732) a045db0 unverified casperhansen winglian commited on Oct 16, 2023
tweak for xformers install w pytorch 2.1.0 (#727) 7f2027d unverified winglian commited on Oct 13, 2023
workaround for installing xformers w torch 2.1.0 (#725) 8d288a2 unverified winglian commited on Oct 13, 2023
update readme to point to direct link to runpod template, cleanup install instrucitons (#532) 34c0a86 unverified winglian commited on Sep 8, 2023
Add support for GPTQ using native transformers/peft (#468) 3355706 unverified winglian commited on Sep 5, 2023
recast loralayer, norm, lmhead + embed token weights per original qlora (#393) 96deb6b unverified winglian commited on Aug 21, 2023
flash attn pip install (#426) cf66547 unverified mhenrichsen Ubuntu mhenrichsen Mads Henrichsen winglian commited on Aug 18, 2023
Merge pull request #108 from OpenAccess-AI-Collective/docker-gptq bbc5bc5 unverified winglian commited on May 30, 2023