Commit History
Feat(cfg): Add code-llama configs for all sizes (#479)
3513071
unverified
Feat(deepspeed): Add zero2 config (#476)
3fc9006
unverified
Feat(doc): Update eval_steps doc (#487)
ad8be43
unverified
Nanobit
commited on
Add example Llama 2 ReLoRA config (#471)
fe4d6ba
unverified
chargoddard
commited on
Merge pull request #486 from OpenAccess-AI-Collective/adam-bnb-simpler
f313010
unverified
tmm1
commited on
let transformers handle adamw_bnb_8bit
868530c
tmm1
commited on
ignore: address pr review
d03887f
unverified
Maxime
commited on
fix: inference did not move the model to the correct device (#483)
17605b8
unverified
Maxime
commited on
ignore: linter
a184549
unverified
Maxime
commited on
fix: finetune model inference needs the dtype fix to work with flash-attn
f311df9
unverified
Maxime
commited on
Fix missing 'packaging' wheel (#482)
c500d02
unverified
Maxime
commited on
fix checkpints on multigpu (#481)
31f3e71
unverified
winglian
commited on
Merge pull request #484 from OpenAccess-AI-Collective/reqs
56c4a94
unverified
tmm1
commited on
allow newer deps
c29117a
tmm1
commited on
fix types w lora (#478)
0b7ba57
unverified
winglian
commited on
Fix(tokenizer): Fix condition to add pad token (#477)
71bd062
unverified
Nanobit
commited on
improve llama pad token handling (#475)
cb9797e
unverified
winglian
commited on
ReLoRA implementation (with quantization) (#322)
bde3c5a
unverified
Fix(doc): Clarify config (#466)
55c23c7
unverified
Nanobit
commited on
workaround so training doesn't hang when packed dataloader batches aren't even (#461)
c69faee
unverified
winglian
commited on
fix test fixture b/c hf trainer tokenization changed (#464)
d5dcf9c
unverified
winglian
commited on
feat: add Metharme prompt strategy (#446)
f474650
unverified
recast loralayer, norm, lmhead + embed token weights per original qlora (#393)
96deb6b
unverified
winglian
commited on
always drop samples that are too long (#452)
50682a3
unverified
winglian
commited on
set env var for FSDP layer to wrap (#453)
5a1985b
unverified
winglian
commited on
Merge pull request #451 from OpenAccess-AI-Collective/eval-is-causal
5e9c6af
unverified
tmm1
commited on
fix eval regression caused in 13f7efaf74fcd3c4514277ccb71914c589873f6a
a213d99
tmm1
commited on
is_causal fix for evals?
fbf49a4
winglian
commited on
add missing positional arg (#450)
58cf7e7
unverified
winglian
commited on
feat(docs): improve user customized prompts (#443)
04a42b6
unverified
Nanobit
commited on
feat(doc): add pillow to lambda instructions (#445)
919f4ca
unverified
Nanobit
commited on
fix evals (#447)
ee26281
unverified
winglian
commited on
gracefully handle empty input (#442)
9d629d8
unverified
winglian
commited on
support user defined prompters, pretokenized datasets in config, local parquet, local arrow files (#348)
d2e7f27
unverified
winglian
commited on
docs(readme): add `cd axolotl` (#440)
d21318d
unverified
philpax
commited on
disable eval using multipack for now (#437)
f733d0f
unverified
winglian
commited on
fix comma, not a tuple (#436)
008505c
unverified
winglian
commited on
use save_strategy from config if available (#434)
b3f5e00
unverified
winglian
commited on
set env for FSDP offload params (#433)
5247c50
unverified
winglian
commited on
flash attn pip install (#426)
cf66547
unverified
adds color (#425)
0a22847
unverified
remove extra accelearate in requirements (#430)
82e111a
unverified
winglian
commited on
fix fixture for new tokenizer handling in transformers (#428)
8cace80
unverified
winglian
commited on
fix orca prompts (#422)
1b7e860
unverified
winglian
commited on
Fix(docs): Remove gptq+lora and fix xformer compat list (#423)
3d1f203
unverified
Nanobit
commited on
just resort to tags ans use main-latest (#424)
d3d6fd6
unverified
winglian
commited on
use inputs for image rather than outputs for docker metadata (#420)
5f80b35
unverified
winglian
commited on