winglian commited on
Commit
fe650dd
1 Parent(s): 49b967b

make sure the CI fails when pytest script fails (#1669)

Browse files

* make sure the pytest script fails

* make sure the defaults come through for tests

* make sure tensorboard is loaded for test assertion

cicd/cicd.sh CHANGED
@@ -1,4 +1,5 @@
1
  #!/bin/bash
 
2
 
3
  pytest --ignore=tests/e2e/ /workspace/axolotl/tests/
4
  pytest /workspace/axolotl/tests/e2e/patched/
 
1
  #!/bin/bash
2
+ set -e
3
 
4
  pytest --ignore=tests/e2e/ /workspace/axolotl/tests/
5
  pytest /workspace/axolotl/tests/e2e/patched/
src/axolotl/utils/samplers/multipack.py CHANGED
@@ -80,8 +80,8 @@ class MultipackBatchSampler(BatchSampler):
80
  self.lengths = np.array(lengths, dtype=np.int32)
81
  self.batch_max_len = batch_max_len
82
  self.batch_size = batch_size
83
- self.group_size = group_size
84
- self.bin_size = bin_size
85
  self.drop_last = drop_last
86
 
87
  self._efficiency = None
 
80
  self.lengths = np.array(lengths, dtype=np.int32)
81
  self.batch_max_len = batch_max_len
82
  self.batch_size = batch_size
83
+ self.group_size = group_size if group_size is not None else 100_000
84
+ self.bin_size = bin_size if bin_size is not None else 200
85
  self.drop_last = drop_last
86
 
87
  self._efficiency = None
tests/e2e/patched/test_resume.py CHANGED
@@ -62,6 +62,7 @@ class TestResumeLlama(unittest.TestCase):
62
  "save_steps": 10,
63
  "save_total_limit": 5,
64
  "max_steps": 40,
 
65
  }
66
  )
67
  if is_torch_bf16_gpu_available():
 
62
  "save_steps": 10,
63
  "save_total_limit": 5,
64
  "max_steps": 40,
65
+ "use_tensorboard": True,
66
  }
67
  )
68
  if is_torch_bf16_gpu_available():