make sure the CI fails when pytest script fails (#1669)
Browse files* make sure the pytest script fails
* make sure the defaults come through for tests
* make sure tensorboard is loaded for test assertion
cicd/cicd.sh
CHANGED
@@ -1,4 +1,5 @@
|
|
1 |
#!/bin/bash
|
|
|
2 |
|
3 |
pytest --ignore=tests/e2e/ /workspace/axolotl/tests/
|
4 |
pytest /workspace/axolotl/tests/e2e/patched/
|
|
|
1 |
#!/bin/bash
|
2 |
+
set -e
|
3 |
|
4 |
pytest --ignore=tests/e2e/ /workspace/axolotl/tests/
|
5 |
pytest /workspace/axolotl/tests/e2e/patched/
|
src/axolotl/utils/samplers/multipack.py
CHANGED
@@ -80,8 +80,8 @@ class MultipackBatchSampler(BatchSampler):
|
|
80 |
self.lengths = np.array(lengths, dtype=np.int32)
|
81 |
self.batch_max_len = batch_max_len
|
82 |
self.batch_size = batch_size
|
83 |
-
self.group_size = group_size
|
84 |
-
self.bin_size = bin_size
|
85 |
self.drop_last = drop_last
|
86 |
|
87 |
self._efficiency = None
|
|
|
80 |
self.lengths = np.array(lengths, dtype=np.int32)
|
81 |
self.batch_max_len = batch_max_len
|
82 |
self.batch_size = batch_size
|
83 |
+
self.group_size = group_size if group_size is not None else 100_000
|
84 |
+
self.bin_size = bin_size if bin_size is not None else 200
|
85 |
self.drop_last = drop_last
|
86 |
|
87 |
self._efficiency = None
|
tests/e2e/patched/test_resume.py
CHANGED
@@ -62,6 +62,7 @@ class TestResumeLlama(unittest.TestCase):
|
|
62 |
"save_steps": 10,
|
63 |
"save_total_limit": 5,
|
64 |
"max_steps": 40,
|
|
|
65 |
}
|
66 |
)
|
67 |
if is_torch_bf16_gpu_available():
|
|
|
62 |
"save_steps": 10,
|
63 |
"save_total_limit": 5,
|
64 |
"max_steps": 40,
|
65 |
+
"use_tensorboard": True,
|
66 |
}
|
67 |
)
|
68 |
if is_torch_bf16_gpu_available():
|