Commit History

hanging slash typo
17345c8

winglian commited on

build on self hosted GPU runners
9cd5d3f

winglian commited on

docker layer caching, build w axolotl from base build
990bec6

winglian commited on

typo in git repo for pip
0c46806

winglian commited on

add huggingface packages and awscli
66fa751

winglian commited on

fix typo and add apex
21b7439

winglian commited on

needs libaio-dev from apt
3f11b47

winglian commited on

pip install packaging dep
ece46b2

winglian commited on

build dependencies and aws-cli
92d800a

winglian commited on

build base separately
2734e3f

winglian commited on

build base too
14ebd2e

winglian commited on

fix push to docker hub
4a79dab

winglian commited on

fix whitespace and instruction on inference
47ad389

winglian commited on

push to docker hub
76b24bc

winglian commited on

TORCH_CUDA_ARCH_LIST should be an ARG
73450d9

winglian commited on

run this on self hosted runner for now
97cf778

winglian commited on

runs on larger git runner?
e2599ed

winglian commited on

don't push the image
75bc856

winglian commited on

run on git commit
15bdbae

winglian commited on

try docker build on gitlab
6603b37

winglian commited on

build dockerfile in gha
2634689

winglian commited on

update stablelm config
4818380

winglian commited on

refactor inference, warn if model is frozen
247825b

winglian commited on

Merge pull request #13 from winglian/dev
cb9a887
unverified

winglian commited on

Merge pull request #12 from NanoCode012/feat/eval_config
a15d823
unverified

winglian commited on

Add eval_batch_size for evaluation
0e74b64

Nanobit commited on

fix log sweep lr
a10a826

winglian commited on

support for multi line inference input, log sweep over learning rates
9105935

winglian commited on

fix adam bnb optimizer grouped parameters, fix peft model 8bit conversion logic, black formatting
7748f3d

winglian commited on

install peft from main branch
fe9c29d

winglian commited on

support llama-adapter zero init attention
2255bb7

winglian commited on

use prebuilt wheels for flash-attn and deepspeed
55baef0

winglian commited on

fdsp config dict fix, todo list, add torchdistx support
ad2b48c

winglian commited on

8bit and deepspeed changes
9190ada

winglian commited on

update ds_config
4dbef09

winglian commited on

don't load models in 8bit unless they are using an adapter, also fix tokenizer load in exceptional case
6dfdd2d

winglian commited on

fix fsdp training args
29936bb

winglian commited on

fix for zero value warmup steps
7882181

winglian commited on

fix sharegpt tokenization, refactor tokenization debugging
5159d00

winglian commited on

wire up gradient checkpointing for 4bit
c0f50d9

winglian commited on

Merge pull request #9 from winglian/dev
4e705ed
unverified

winglian commited on

fix dataset handling, support galactica
4a17a4c

winglian commited on

tweaks to data loading, 8 bit adam, accelerate and deepspeed
097d367

winglian commited on

shuffle and split dataset after save/load
4f2584f

winglian commited on

fix sharegpt handling from hf, don't worry about loading llama if using earlier transformers release
8d43785

winglian commited on

stablelm support
8e2a560

winglian commited on

various bugfixes
94f5e41

winglian commited on

ignore config, add python 3.9 (#8)
2624bc2
unverified

ehartford commited on

fix bug when model_type not explicitly passed
bb991fd

winglian commited on

improve inference
d653859

winglian commited on