Commit History
skip some flash attn patches unless explicitly enabled (#643)
895f0a0
unverified
winglian
commited on
update for recent transformers updates (#636)
60c7c48
unverified
winglian
commited on
Add training callback to send predictions to WandB table (#521)
5b67ea9
unverified
Glavin001
commited on
reorg a bit
fc8766e
tmm1
commited on
use flash_attn rmsnorm when available (#526)
72a6fe1
unverified
tmm1
commited on
use flash_attn xentropy when available (#525)
5fe30b1
unverified
tmm1
commited on
fix eval regression caused in 13f7efaf74fcd3c4514277ccb71914c589873f6a
a213d99
tmm1
commited on
is_causal fix for evals?
fbf49a4
winglian
commited on
fix evals (#447)
ee26281
unverified
winglian
commited on
fix check for flash attn branching (#377)
343ac84
unverified
winglian
commited on
Attention mask and position id fixes for packing (#285)
2bb0b78
unverified
winglian
commited on
move flash-attn monkey patch alongside the others
312a9fa
tmm1
commited on