Commit History
optionally be able to specify alpaca or chat style prompts
1d5ab84
winglian
commited on
add alpaca multiple choice instruct dataset support
b46bc02
winglian
commited on
reorder options so debug can happen in the same prepare step
f98e173
winglian
commited on
more fixes
bdbca8f
winglian
commited on
move filter to before saving so it doesn't happen everytime, update runpod manual script
0d28df0
winglian
commited on
Fix typo
52aada7
unverified
Nanobit
commited on
black formatting
2bc1a5b
winglian
commited on
Update finetune.py
915c56c
unverified
winglian
commited on
Don't save full model for lora
cd23959
unverified
Nanobit
commited on
Save adapter for lora
71a1f7f
unverified
Nanobit
commited on
setup runpod images
79deb35
winglian
commited on
fix whitespace and instruction on inference
47ad389
winglian
commited on
refactor inference, warn if model is frozen
247825b
winglian
commited on
support for multi line inference input, log sweep over learning rates
9105935
winglian
commited on
install peft from main branch
fe9c29d
winglian
commited on
support llama-adapter zero init attention
2255bb7
winglian
commited on
use prebuilt wheels for flash-attn and deepspeed
55baef0
winglian
commited on
fix sharegpt tokenization, refactor tokenization debugging
5159d00
winglian
commited on
various bugfixes
94f5e41
winglian
commited on
improve inference
d653859
winglian
commited on
fix runpod script
5749eb0
winglian
commited on
quickstart instructions for starting from runpod (#5)
0a472e1
unverified
winglian
commited on
WIP large refactor to make finetune script a little more manageable (#3)
6045345
unverified
winglian
commited on
add support for alpaca reflect training (#2)
81de0ef
unverified
winglian
commited on
Tokenization open assistant (#1)
87d7825
unverified
winglian
commited on
fix llama check
eb80890
winglian
commited on
fix conditional check to prevent always using 4bit
8f36f3c
winglian
commited on
imrpove llama check and fix safetensors file check
69164da
winglian
commited on
suppport for alpaca-like instruction datasets without inputs
e107643
winglian
commited on
casts the prepared data to int16 (doesn't help with training memory)
2db9436
winglian
commited on
bugfixes
120e7df
winglian
commited on
fix lora target module, require explicit flash attention, fix min logging steps, don't use adam8bit for int4, hash prepared datasets, support hf hub datasets
87e073d
winglian
commited on
4bit quantized support (wip)
77fca25
winglian
commited on
cleanup, prep for 4bit quant support
12de7b7
winglian
commited on
deepspeed doesn't work with flash-attn, and the gpu savings w flash attn are better than the deepspeed headaches
d1aed4c
winglian
commited on
fix logging
a459383
winglian
commited on
prepare datasets only flag
2393801
winglian
commited on
configure log level, add llama 7b config
d33a975
winglian
commited on
more logging, wandb fixes
05fffb5
winglian
commited on
refactor trainer setup to account for deepspeed integration
2df63ef
winglian
commited on
improve prepared dataset loading, fix inference
b164725
winglian
commited on
helpful info output
937f44f
winglian
commited on
fix issue with completed model being empty
902dd0a
winglian
commited on
various bugfixes
80b2ed2
winglian
commited on
bettter handling of llama model import
45f77dd
winglian
commited on
more fixes and prep for llama training
949a27b
winglian
commited on
config chooser, update readme instructions, device config, llama flash attention, debug out the labels, fix config key checks, other bugfixes
f2a2029
winglian
commited on
black formatting
a6028d3
winglian
commited on