* add more test cases for gradient accumulation and fix zero3 * swap out for smaller model * fix missing return * fix missing pad_token in config * support concurrency for multigpu testing * cast empty deepspeed to empty string for zero3 check * fix temp_dir as fixture so parametrize works properly * fix test file for multigpu evals * don't use default * don't use default for fsdp_state_dict_type * don't use llama tokenizer w smollm * also automatically cancel multigpu for concurrency
6 lines
128 B
Bash
Executable File
6 lines
128 B
Bash
Executable File
#!/bin/bash
|
|
set -e
|
|
|
|
# only run one test at a time so as not to OOM the GPU
|
|
pytest -v -n2 /workspace/axolotl/tests/e2e/multigpu/
|