optimize dataloading to use cache, fix model token embedding sizes aa3c3f9 winglian commited on May 12, 2023
Fix Trainer() got multiple values for keyword argument 'callbacks' 813aab3 unverified Nanobit commited on May 10, 2023
Merge pull request #19 from NanoCode012/feat/callback-save-lora bcbc99e unverified winglian commited on May 8, 2023
support for multi line inference input, log sweep over learning rates 9105935 winglian commited on May 3, 2023
fix adam bnb optimizer grouped parameters, fix peft model 8bit conversion logic, black formatting 7748f3d winglian commited on May 1, 2023
don't load models in 8bit unless they are using an adapter, also fix tokenizer load in exceptional case 6dfdd2d winglian commited on Apr 30, 2023
tweaks to data loading, 8 bit adam, accelerate and deepspeed 097d367 winglian commited on Apr 22, 2023
fix sharegpt handling from hf, don't worry about loading llama if using earlier transformers release 8d43785 winglian commited on Apr 20, 2023
quickstart instructions for starting from runpod (#5) 0a472e1 unverified winglian commited on Apr 18, 2023
WIP large refactor to make finetune script a little more manageable (#3) 6045345 unverified winglian commited on Apr 18, 2023
suppport for alpaca-like instruction datasets without inputs e107643 winglian commited on Apr 18, 2023
casts the prepared data to int16 (doesn't help with training memory) 2db9436 winglian commited on Apr 18, 2023
config chooser, update readme instructions, device config, llama flash attention, debug out the labels, fix config key checks, other bugfixes f2a2029 winglian commited on Apr 14, 2023