qwerrwe / requirements.txt

Commit History

ORPO Trainer replacement (#1551)
7d1d22f
unverified

winglian commited on

fix(packages): lock datasets version (#1545)
59ef254
unverified

Nanobit commited on

DBRX Model Support (#1462)
132eb74
unverified

winglian commited on

Pretrain multipack v2 (#1470)
5aa5097
unverified

winglian commited on

qwen2_moe support w multipack (#1455)
6086be8
unverified

winglian commited on

fix some of the edge cases for Jamba (#1452)
05b398a
unverified

winglian commited on

strip out hacky qlora-fsdp workarounds now that qlora-fsdp fixes are upstreamed (#1428)
2a1589f
unverified

winglian commited on

support galore once upstreamed into transformers (#1409)
dd449c5
unverified

winglian commited on

FDSP + QLoRA (#1378)
9b6ee83
unverified

winglian commited on

update flash attention for gemma support: (#1368)
58b0d4b
unverified

winglian commited on

support for DoRA w/ PEFT (#1363)
0cfdb2c
unverified

winglian commited on

run tests again on Modal (#1289) [skip ci]
0001862
unverified

winglian commited on

fix: checkpoint saving with deepspeed (#1321)
5be8b55
unverified

Nanobit commited on

Pydantic 2.x cfg (#1239)
cc3cebf
unverified

winglian commited on

make mlflow optional (#1317)
5894f0e
unverified

winglian commited on

multipack for gemma (#1313)
2752d5f
unverified

winglian commited on

Add seq2seq eval benchmark callback (#1274)
5a5d474
unverified

LeonardoEmili commited on

add support for https remote yamls (#1277)
9bca7db
unverified

hamel commited on

Peft deepspeed resume (#1227)
c67fb71
unverified

winglian commited on

Peft lotfq (#1222)
4cb7900
unverified

winglian commited on

Revert "run PR e2e docker CI tests in Modal" (#1220) [skip ci]
8da1633
unverified

winglian commited on

run PR e2e docker CI tests in Modal (#1217) [skip ci]
36d053f
unverified

winglian commited on

Update deps 202401 (#1204) [skip ci]
a01b998
unverified

winglian commited on

upgrade deepspeed to 0.13.1 for mixtral fixes (#1189) [skip ci]
8a49309
unverified

winglian commited on

Qwen2 (#1166)
f5a828a
unverified

winglian commited on

Remove fused-dense-lib from requirements.txt (#1087)
91502b9
unverified

casperhansen commited on

fix: warn user to install mamba_ssm package (#1019)
d69ba2b
unverified

Nanobit commited on

pin accelerate for deepspeed fix (#1080)
9e3f0cb
unverified

winglian commited on

Separate AutoGPTQ dep to `pip install -e .[auto-gptq]` (#1077)
9be92d1
unverified

casperhansen commited on

paired kto support (#1069)
d7057cc
unverified

winglian commited on

update peft to 0.7.0 (#1073)
768d348
unverified

marktenenholtz commited on

Add: mlflow for experiment tracking (#1059) [skip ci]
090c24d
unverified

Johan Hansson winglian commited on

Phi2 rewrite (#1058)
732851f
unverified

winglian commited on

RL/DPO (#935)
f243c21

winglian commited on

bump transformers and update attention class map name (#1023)
bcc78d8
unverified

winglian commited on

chore: Update transformers to latest (#986)
7d4185f
unverified

Nanobit commited on

update transformers to fix checkpoint saving (#963)
f28e755
unverified

dumpmemory commited on

Mixtral official (#942)
7fabc4d
unverified

winglian commited on

Update requirements.txt (#940)
9a5eb39
unverified

tokestermw commited on

update to latest transformers for mixstral support (#929)
35f9b0f
unverified

winglian commited on

update datasets version to cut down the warnings due to pyarrow arg change (#897)
6a4562a
unverified

winglian commited on

try #2: pin hf transformers and accelerate to latest release, don't reinstall pytorch (#867)
0de1457
unverified

winglian commited on

Feat: Add dataset loading from S3, GCS (#765)
3cc67d2
unverified

Nanobit commited on

add e2e tests for checking functionality of resume from checkpoint (#865)
b3a61e8
unverified

winglian commited on

Pin optimum package (#838)
105d0b3
unverified

Bryan Thornbury commited on

don't compile deepspeed or bitsandbytes from source (#837)
f544ab2
unverified

winglian commited on

Feat: Added Gradio support (#812)
738a057
unverified

stillerman commited on

fix: pin autogptq (#818)
6459ac7
unverified

Nanobit commited on

chore: bump transformers to v4.34.1 to fix tokenizer issue (#745)
8966a6f
unverified

Nanobit commited on