Commit History

FIX: max_length and max_prompt_length was not being sent to ORPOTrainer (#1584)
1e1921b
unverified

alimosavian Ali Mosavian winglian commited on

improve save callbacks (#1592)
29cf15a
unverified

winglian commited on

FIX: TRL trainer preprocessing step was running in one process (#1583)
b9bb169
unverified

Ali Mosavian Ali Mosavian commited on

PoSE context length ext (#1567)
5294653
unverified

winglian commited on

make sure everything stays in the same dtype when using dpo + FSDP (#1559)
68601ec
unverified

winglian commited on

ORPO Trainer replacement (#1551)
7d1d22f
unverified

winglian commited on

DBRX Model Support (#1462)
132eb74
unverified

winglian commited on

WIP: Support table logging for mlflow, too (#1506)
057fa44
unverified

DavidFarago Dave Farago winglian commited on

drop empty token from beginning if tokenizer has no bos_token (in the case of qwen) (#1490)
934fc85
unverified

winglian commited on

Fix ORPO multi gpu (#1433)
34ba634
unverified

winglian commited on

strip out hacky qlora-fsdp workarounds now that qlora-fsdp fixes are upstreamed (#1428)
2a1589f
unverified

winglian commited on

HF / FEAT: Optimize HF tags (#1425) [skip ci]
7d55607
unverified

Younes Belkada winglian commited on

support galore once upstreamed into transformers (#1409)
dd449c5
unverified

winglian commited on

fix(config): passing gradient_checkpoint_kwargs (#1412)
b1e3e1b
unverified

Nanobit commited on

ORPO (#1419)
2ea70eb
unverified

winglian commited on

FDSP + QLoRA (#1378)
9b6ee83
unverified

winglian commited on

lora+ support (#1352)
decb66e
unverified

winglian commited on

add lion-pytorch optimizer (#1299) [skip ci]
1648279
unverified

Maxime winglian commited on

make mlflow optional (#1317)
5894f0e
unverified

winglian commited on

Allow load_best_model_at_end to be configured for early stopping on custom evaluation datasets (#1291)
3c00f40
unverified

David Meikle commited on

Add seq2seq eval benchmark callback (#1274)
5a5d474
unverified

LeonardoEmili commited on

Scheduler implementation of Continual Pre-Training of Large Language Models: How to (re)warm your model? (#1273)
8430db2
unverified

jinwonkim93 commited on

allow the optimizer prune ratio for ReLoRA to be configurable (#1287)
4b997c3
unverified

winglian commited on

simplify haldning for newer multipack patches so they can be added in a single place (#1270)
5698943
unverified

winglian commited on

Add more save strategies for DPO training. (#1255)
13eea21
unverified

Philip May commited on

relora: magnitude pruning of the optimizer (#1245)
8c2e05a
unverified

winglian commited on

support for true batches with multipack (#1230)
00568c1
unverified

winglian commited on

Fix and document test_datasets (#1228)
5787e1a
unverified

DreamGenX winglian commited on

FEAT: add tagging support to axolotl for DPOTrainer (#1209)
18f8119
unverified

Filippo Broggini winglian commited on

precompute dpo logprobs setting and fixes (#1199) [skip ci]
33e1170
unverified

winglian commited on

fix learning rate scheduler's warnings (#1135) [skip ci]
b4ac96a
unverified

ricdomolm winglian commited on

more dpo fixes for dataset loading and docs (#1185) [skip ci]
5bce45f
unverified

winglian commited on

DPO fixes v2 (#1174)
59a31fe
unverified

winglian commited on

Phi2 multipack (#1173)
814aee6
unverified

winglian commited on

Add mlflow callback for pushing config to mlflow artifacts (#1125)
b8e5603
unverified

JohanWork commited on

jupyter lab fixes (#1139) [skip ci]
eaaeefc
unverified

winglian commited on

Qwen2 (#1166)
f5a828a
unverified

winglian commited on

Multipack simplify for Mixtral (#1142)
6910e6a
unverified

winglian commited on

swap the data collator for evals if not using sample packing (#1076)
ead34c5
unverified

winglian commited on

paired kto support (#1069)
d7057cc
unverified

winglian commited on

Add: mlflow for experiment tracking (#1059) [skip ci]
090c24d
unverified

Johan Hansson winglian commited on

Cosine learning rate schedule - minimum learning rate (#1062)
04b978b
unverified

ricdomolm winglian commited on

Efficiently get the length of the tokenized docs (#1063)
81d3845
unverified

ricdomolm winglian commited on

Phi2 rewrite (#1058)
732851f
unverified

winglian commited on

streaming multipack for pretraining dataset (#959)
553c80f
unverified

jinwonkim93 jinwonkim93@github.com winglian commited on

feat: always push checkpoint to hub if set (#1049) [skip ci]
cbdbf9e
unverified

Nanobit commited on

RL/DPO (#935)
f243c21

winglian commited on

use recommended setting for use_reentrant w gradient checkpointing (#1021)
4d2e842
unverified

winglian commited on