Commit History
improve save callbacks (#1592)
29cf15a
unverified
winglian
commited on
FIX: TRL trainer preprocessing step was running in one process (#1583)
b9bb169
unverified
Ali Mosavian
Ali Mosavian
commited on
PoSE context length ext (#1567)
5294653
unverified
winglian
commited on
make sure everything stays in the same dtype when using dpo + FSDP (#1559)
68601ec
unverified
winglian
commited on
ORPO Trainer replacement (#1551)
7d1d22f
unverified
winglian
commited on
DBRX Model Support (#1462)
132eb74
unverified
winglian
commited on
WIP: Support table logging for mlflow, too (#1506)
057fa44
unverified
drop empty token from beginning if tokenizer has no bos_token (in the case of qwen) (#1490)
934fc85
unverified
winglian
commited on
Fix ORPO multi gpu (#1433)
34ba634
unverified
winglian
commited on
strip out hacky qlora-fsdp workarounds now that qlora-fsdp fixes are upstreamed (#1428)
2a1589f
unverified
winglian
commited on
HF / FEAT: Optimize HF tags (#1425) [skip ci]
7d55607
unverified
support galore once upstreamed into transformers (#1409)
dd449c5
unverified
winglian
commited on
fix(config): passing gradient_checkpoint_kwargs (#1412)
b1e3e1b
unverified
Nanobit
commited on
ORPO (#1419)
2ea70eb
unverified
winglian
commited on
FDSP + QLoRA (#1378)
9b6ee83
unverified
winglian
commited on
lora+ support (#1352)
decb66e
unverified
winglian
commited on
add lion-pytorch optimizer (#1299) [skip ci]
1648279
unverified
make mlflow optional (#1317)
5894f0e
unverified
winglian
commited on
Allow load_best_model_at_end to be configured for early stopping on custom evaluation datasets (#1291)
3c00f40
unverified
David Meikle
commited on
Add seq2seq eval benchmark callback (#1274)
5a5d474
unverified
LeonardoEmili
commited on
Scheduler implementation of Continual Pre-Training of Large Language Models: How to (re)warm your model? (#1273)
8430db2
unverified
jinwonkim93
commited on
allow the optimizer prune ratio for ReLoRA to be configurable (#1287)
4b997c3
unverified
winglian
commited on
simplify haldning for newer multipack patches so they can be added in a single place (#1270)
5698943
unverified
winglian
commited on
Add more save strategies for DPO training. (#1255)
13eea21
unverified
Philip May
commited on
relora: magnitude pruning of the optimizer (#1245)
8c2e05a
unverified
winglian
commited on
support for true batches with multipack (#1230)
00568c1
unverified
winglian
commited on
Fix and document test_datasets (#1228)
5787e1a
unverified
FEAT: add tagging support to axolotl for DPOTrainer (#1209)
18f8119
unverified
precompute dpo logprobs setting and fixes (#1199) [skip ci]
33e1170
unverified
winglian
commited on
more dpo fixes for dataset loading and docs (#1185) [skip ci]
5bce45f
unverified
winglian
commited on
DPO fixes v2 (#1174)
59a31fe
unverified
winglian
commited on
Phi2 multipack (#1173)
814aee6
unverified
winglian
commited on
Add mlflow callback for pushing config to mlflow artifacts (#1125)
b8e5603
unverified
JohanWork
commited on
jupyter lab fixes (#1139) [skip ci]
eaaeefc
unverified
winglian
commited on
Qwen2 (#1166)
f5a828a
unverified
winglian
commited on
Multipack simplify for Mixtral (#1142)
6910e6a
unverified
winglian
commited on
swap the data collator for evals if not using sample packing (#1076)
ead34c5
unverified
winglian
commited on
paired kto support (#1069)
d7057cc
unverified
winglian
commited on
Add: mlflow for experiment tracking (#1059) [skip ci]
090c24d
unverified
Phi2 rewrite (#1058)
732851f
unverified
winglian
commited on
streaming multipack for pretraining dataset (#959)
553c80f
unverified
feat: always push checkpoint to hub if set (#1049) [skip ci]
cbdbf9e
unverified
Nanobit
commited on
RL/DPO (#935)
f243c21
winglian
commited on
use recommended setting for use_reentrant w gradient checkpointing (#1021)
4d2e842
unverified
winglian
commited on