make sure the CI fails when pytest script fails (#1669)
Browse files* make sure the pytest script fails
* make sure the defaults come through for tests
* make sure tensorboard is loaded for test assertion
cicd/cicd.sh
CHANGED
|
@@ -1,4 +1,5 @@
|
|
| 1 |
#!/bin/bash
|
|
|
|
| 2 |
|
| 3 |
pytest --ignore=tests/e2e/ /workspace/axolotl/tests/
|
| 4 |
pytest /workspace/axolotl/tests/e2e/patched/
|
|
|
|
| 1 |
#!/bin/bash
|
| 2 |
+
set -e
|
| 3 |
|
| 4 |
pytest --ignore=tests/e2e/ /workspace/axolotl/tests/
|
| 5 |
pytest /workspace/axolotl/tests/e2e/patched/
|
src/axolotl/utils/samplers/multipack.py
CHANGED
|
@@ -80,8 +80,8 @@ class MultipackBatchSampler(BatchSampler):
|
|
| 80 |
self.lengths = np.array(lengths, dtype=np.int32)
|
| 81 |
self.batch_max_len = batch_max_len
|
| 82 |
self.batch_size = batch_size
|
| 83 |
-
self.group_size = group_size
|
| 84 |
-
self.bin_size = bin_size
|
| 85 |
self.drop_last = drop_last
|
| 86 |
|
| 87 |
self._efficiency = None
|
|
|
|
| 80 |
self.lengths = np.array(lengths, dtype=np.int32)
|
| 81 |
self.batch_max_len = batch_max_len
|
| 82 |
self.batch_size = batch_size
|
| 83 |
+
self.group_size = group_size if group_size is not None else 100_000
|
| 84 |
+
self.bin_size = bin_size if bin_size is not None else 200
|
| 85 |
self.drop_last = drop_last
|
| 86 |
|
| 87 |
self._efficiency = None
|
tests/e2e/patched/test_resume.py
CHANGED
|
@@ -62,6 +62,7 @@ class TestResumeLlama(unittest.TestCase):
|
|
| 62 |
"save_steps": 10,
|
| 63 |
"save_total_limit": 5,
|
| 64 |
"max_steps": 40,
|
|
|
|
| 65 |
}
|
| 66 |
)
|
| 67 |
if is_torch_bf16_gpu_available():
|
|
|
|
| 62 |
"save_steps": 10,
|
| 63 |
"save_total_limit": 5,
|
| 64 |
"max_steps": 40,
|
| 65 |
+
"use_tensorboard": True,
|
| 66 |
}
|
| 67 |
)
|
| 68 |
if is_torch_bf16_gpu_available():
|