drop length column for issues with eval without packing (#1711) 3f1f5e3 unverified winglian commited on Jun 19
fix for when sample_packing and eval_sample_packing are different (#1695) 18cabc0 unverified winglian commited on Jun 8
add back packing efficiency estimate so epochs and multi-gpu works properly (#1697) ed8ef65 unverified winglian commited on Jun 8
ensure explicit eval_sample_packing to avoid mismatch issues (#1692) 9c1af1a unverified winglian commited on Jun 7
Phi-3 conversation format, example training script and perplexity metric (#1582) cf64284 unverified roborovski winglian commited on Jun 4
cleanup the deepspeed proxy model at the end of training (#1675) d4f6c65 unverified winglian commited on May 30
set chat_template in datasets config automatically (#1664) 9d4225a unverified winglian commited on May 30
use mixins for orpo and kto configs so they work with axolotl customizations (#1674) f7332ac unverified winglian commited on May 30
make sure the CI fails when pytest script fails (#1669) fe650dd unverified winglian commited on May 29
Correct name of MixtralBlockSparseTop2MLP (L -> l) (#1667) 65db903 unverified seungduk commited on May 28
Fix: ensure correct handling of `val_set_size` as `float` or `int` (#1655) 6a5a725 unverified Davide Caroselli winglian commited on May 28
Generalizing the chat_template prompt strategy (#1660) [skip ci] cc11c6b unverified fozziethebeat commited on May 28
Switch to parallel FFD bin packing algorithm. (#1619) 367b2e8 unverified winglian daaave commited on May 23
enable loraplus setting for dpo trainer (#1646) a27d5e1 unverified thepowerfuldeez commited on May 22
Fix llama3 chat_template (extra <|eot_id|> on last turn) (#1635) 7c2bf30 unverified leonardlin winglian commited on May 21
FIX: max_length and max_prompt_length was not being sent to ORPOTrainer (#1584) 1e1921b unverified alimosavian Ali Mosavian winglian commited on May 14
feat: Add LLaMA-3 instruct prompt strategies for fine-tuning (#1553) 50421c8 unverified Ram Ram winglian commited on May 11
adding llama3 fastchat conversation monkeypatch (#1539) b32c08f unverified Antoni-Joan Solergibert winglian commited on May 10
make sure to save the lora adapter at the end of RL/dpo training (#1573) 796a085 unverified winglian commited on May 8
Pass deepspeed and fsdp as None explicitly when merging adapters to allow custom device_map (#1575) 9e1480e unverified chiragjn commited on May 7
Gradio configuration parameters (#1591) 3367fca unverified marijnfs Marijn Stollenga Marijn Stollenga winglian commited on May 6
Pass weakref to model in the SIGINT handler to free up model post train function (#1581) dde02fc unverified chiragjn winglian commited on May 3
FIX: TRL trainer preprocessing step was running in one process (#1583) b9bb169 unverified Ali Mosavian Ali Mosavian commited on May 3
Add debug option for RL dataset preprocessing (#1404) cc5d31e unverified abhinand Nanobit commited on Apr 30
make sure everything stays in the same dtype when using dpo + FSDP (#1559) 68601ec unverified winglian commited on Apr 22
Add support for Gemma chat template (#1530) 60f5ce0 unverified Haoxiang-Wang winglian commited on Apr 21