move is_llama_derived_model into normalize_config (#524) 44454ae unverified tmm1 commited on Sep 4, 2023
Debug tokenization output: Add ability to output text only (no tokens), and/or specify num samples to see (#511) 48434be unverified Tom Jobbins commited on Aug 31, 2023
tweak: use default config file when only one file is present (#501) 36b2e1c unverified Maxime commited on Aug 29, 2023
fix: inference did not move the model to the correct device (#483) 17605b8 unverified Maxime commited on Aug 26, 2023
ReLoRA implementation (with quantization) (#322) bde3c5a unverified chargoddard winglian commited on Aug 24, 2023
use context manager to run things on rank0 before others (#397) fc2d6be unverified winglian commited on Aug 15, 2023
Attention mask and position id fixes for packing (#285) 2bb0b78 unverified winglian commited on Aug 12, 2023
Fixed pre-commit problems, fixed small bug in logging_config to handle LOG_LEVEL env var b1f4f7a theobjectivedad commited on Jul 15, 2023
Merge pull request #92 from OpenAccess-AI-Collective/flash-optimum 16bb627 unverified winglian commited on Jun 14, 2023
Merge pull request #177 from NanoCode012/fix/landmark-patch 8002ffb unverified winglian commited on Jun 12, 2023
add flash attn context for efficient training and attempt setting model to train mode: 8792199 winglian commited on May 27, 2023
Remove explicit definition of cfg.inference c250898 unverified Angainor Development commited on Jun 10, 2023
new prompters, misc fixes for output dir missing using fsdp, and changing max seq len 4ac9e25 winglian commited on Jun 6, 2023