Commit History
adds color (#425)
0a22847
unverified
remove extra accelearate in requirements (#430)
82e111a
unverified
winglian
commited on
fix fixture for new tokenizer handling in transformers (#428)
8cace80
unverified
winglian
commited on
fix orca prompts (#422)
1b7e860
unverified
winglian
commited on
Fix(docs): Remove gptq+lora and fix xformer compat list (#423)
3d1f203
unverified
Nanobit
commited on
just resort to tags ans use main-latest (#424)
d3d6fd6
unverified
winglian
commited on
use inputs for image rather than outputs for docker metadata (#420)
5f80b35
unverified
winglian
commited on
hopefully improve the README (#419)
2495909
unverified
winglian
commited on
tag with latest as well for axolotl-runpod (#418)
7af8166
unverified
winglian
commited on
Merge pull request #413 from mhenrichsen/chore/update-deepseed-config
f806e86
unverified
mhenrichsen
commited on
Feat(doc): Add lr_quadratic_warmup to readme (#412)
2b990eb
unverified
Nanobit
commited on
update path to align with fsdp example
bd8cab4
mhenrichsen
commited on
Fix(config): Update handling of deepspeed config (#404)
c01015f
unverified
Nanobit
commited on
Fix(docs): Update flash attn requirements (#409)
72fe3f8
unverified
Nanobit
commited on
update docs for tokenizer_legacy (#401)
47961fd
unverified
winglian
commited on
Fix(template): Remove iPhone/android from Issue template (#407)
7ad37cb
unverified
Nanobit
commited on
Ax art (#405)
29241cf
unverified
winglian
commited on
fix eval steps and strategy (#403)
da10af0
unverified
winglian
commited on
better handling of empty input ids when tokenizing (#395)
85cf4f8
unverified
winglian
commited on
add utils.data.prepare_dataset
2e22404
tmm1
commited on
Feat(doc): Add how to save by epochs (#396)
be294fd
unverified
Nanobit
commited on
use context manager to run things on rank0 before others (#397)
fc2d6be
unverified
winglian
commited on
don't use mask expansion for inference (#392)
1687be6
unverified
winglian
commited on
Feat(doc): Add max_steps to readme (#389)
41ecb45
unverified
Nanobit
commited on
Feat(config): add max steps (#387)
3c2ad00
unverified
ittailup
commited on
Added "epoch" evaluation_strategy (#388)
5d48a10
unverified
flotos
commited on
Feat(config): Add hub_strategy (#386)
73a0b6e
unverified
Nanobit
commited on
Error msg for sharegpt if conv has less than 2 msg (#379)
63fdb5a
unverified
flotos
commited on
new llama-2 default settings (#370)
fdffef5
unverified
don't pass rope_scaling kwarg if it's None (#383)
919246f
unverified
winglian
commited on
bump flash-attn to 2.0.4 for the base docker image (#382)
ffac902
unverified
winglian
commited on
Fix crash when running without CUDA
15f6e57
chargoddard
commited on
Feat(doc): Improve sharegpt doc (#378)
729c299
unverified
Nanobit
commited on
save tokenizer before training starts (#380)
86a91e2
unverified
winglian
commited on
try to detect accelerate and only use device_map=None in that case (#373)
094fc2c
unverified
tmm1
commited on
Create FUNDING.yml
2dafa73
unverified
winglian
commited on
fix check for flash attn branching (#377)
343ac84
unverified
winglian
commited on
remove unnecessary local variable
0c96727
tmm1
commited on
simplify `load_tokenizer`
efb3b2c
tmm1
commited on
improve GPU logging to break out pytorch cache and system mem
7b55fe6
tmm1
commited on
quiet noise from llama tokenizer by setting pad token earlier
e029ab3
tmm1
commited on
extract module for working with cfg
8cec513
tmm1
commited on
fix DefaultDict.__or__
a13e45d
tmm1
commited on
revert previous change and build ax images w docker on gpu (#371)
918f1b0
unverified
winglian
commited on
attempt to run non-base docker builds on regular cpu hosts (#369)
c3fde36
unverified
winglian
commited on
Attention mask and position id fixes for packing (#285)
2bb0b78
unverified
winglian
commited on
Fix(save): Save as safetensors (#363)
a276c9c
unverified
Nanobit
commited on