Compare commits
16 Commits
fa-check
...
fix-previe
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
7d8e8c9ac2 | ||
|
|
7c2466b739 | ||
|
|
3146cb56dd | ||
|
|
c09b0a3bbf | ||
|
|
e05acccd77 | ||
|
|
c44abad531 | ||
|
|
817d70e669 | ||
|
|
03f5a7fd16 | ||
|
|
3d9b96a94f | ||
|
|
42c16024a2 | ||
|
|
ec94d632f3 | ||
|
|
e8bd3b0b3b | ||
|
|
5a08b94668 | ||
|
|
ecb8c1f4b3 | ||
|
|
ab57be6526 | ||
|
|
c9640bca2c |
7
.github/CONTRIBUTING.md
vendored
7
.github/CONTRIBUTING.md
vendored
@@ -57,6 +57,13 @@ We welcome ideas for improvements and new features. To suggest an enhancement, o
|
|||||||
5. Push your branch to your fork on GitHub.
|
5. Push your branch to your fork on GitHub.
|
||||||
6. Open a new pull request against the `main` branch of the axolotl repository. Include a clear and concise description of your changes, referencing any related issues.
|
6. Open a new pull request against the `main` branch of the axolotl repository. Include a clear and concise description of your changes, referencing any related issues.
|
||||||
|
|
||||||
|
#### Skipping CI Checks
|
||||||
|
|
||||||
|
You can skip certain CI checks by including specific keywords in your commit messages:
|
||||||
|
|
||||||
|
- `[skip ci]` or `skip ci` - Skips all CI checks for that commit
|
||||||
|
- `[skip-e2e]` or `skip-e2e` - Skips only end-to-end tests while running other CI checks
|
||||||
|
|
||||||
## Style Guidelines
|
## Style Guidelines
|
||||||
|
|
||||||
### Code Style
|
### Code Style
|
||||||
|
|||||||
10
TODO.md
10
TODO.md
@@ -1,10 +0,0 @@
|
|||||||
# todo list
|
|
||||||
|
|
||||||
- [] Validation of parameters for combinations that won't work
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
## things that are known not to work
|
|
||||||
|
|
||||||
- FSDP offload and gradient_checkpointing - https://github.com/pytorch/pytorch/issues/82203
|
|
||||||
- adamw_bnb_8bit doesn't play well with FSDP offload
|
|
||||||
@@ -14,7 +14,7 @@ packaging==23.2
|
|||||||
|
|
||||||
huggingface_hub>=0.33.0
|
huggingface_hub>=0.33.0
|
||||||
peft==0.17.0
|
peft==0.17.0
|
||||||
transformers @ git+https://github.com/vasqu/transformers@fix-fa-integration
|
transformers==4.55.0
|
||||||
tokenizers>=0.21.1
|
tokenizers>=0.21.1
|
||||||
accelerate==1.10.0
|
accelerate==1.10.0
|
||||||
datasets==4.0.0
|
datasets==4.0.0
|
||||||
|
|||||||
@@ -67,14 +67,12 @@ def build_command(base_cmd: list[str], options: dict[str, Any]) -> list[str]:
|
|||||||
|
|
||||||
def generate_config_files(config: str, sweep: str | None) -> Iterator[tuple[str, bool]]:
|
def generate_config_files(config: str, sweep: str | None) -> Iterator[tuple[str, bool]]:
|
||||||
"""
|
"""
|
||||||
Generate list of configuration files to process.
|
Generate list of configuration files to process. Yields a tuple of the configuration file name and a boolean indicating
|
||||||
|
whether this is a group of configurations (i.e., a sweep).
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
config: Base configuration file
|
config: Base configuration file
|
||||||
sweep: Sweep configuration file
|
sweep: Sweep configuration file
|
||||||
|
|
||||||
Yields:
|
|
||||||
Tuple of configuration file name and whether this is a group of configurations
|
|
||||||
"""
|
"""
|
||||||
|
|
||||||
if not sweep:
|
if not sweep:
|
||||||
|
|||||||
@@ -18,7 +18,9 @@ from torch.distributed import DeviceMesh
|
|||||||
try:
|
try:
|
||||||
from transformers.modeling_flash_attention_utils import _flash_supports_window
|
from transformers.modeling_flash_attention_utils import _flash_supports_window
|
||||||
except ImportError:
|
except ImportError:
|
||||||
_flash_supports_window = True
|
from transformers.modeling_flash_attention_utils import (
|
||||||
|
_flash_supports_window_size as _flash_supports_window,
|
||||||
|
)
|
||||||
|
|
||||||
from axolotl.monkeypatch.utils import get_cu_seqlens_from_pos_ids
|
from axolotl.monkeypatch.utils import get_cu_seqlens_from_pos_ids
|
||||||
from axolotl.utils.logging import get_logger
|
from axolotl.utils.logging import get_logger
|
||||||
|
|||||||
@@ -3,6 +3,7 @@
|
|||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
from axolotl.monkeypatch.transformers.trainer_loss_calc import (
|
from axolotl.monkeypatch.transformers.trainer_loss_calc import (
|
||||||
|
check_evaluation_loop_is_fsdp2_patchable,
|
||||||
check_evaluation_loop_is_patchable,
|
check_evaluation_loop_is_patchable,
|
||||||
check_maybe_log_save_evaluate_is_patchable,
|
check_maybe_log_save_evaluate_is_patchable,
|
||||||
)
|
)
|
||||||
@@ -19,6 +20,7 @@ class TestTrainerLossCalc(unittest.TestCase):
|
|||||||
the patched code changes upstream.
|
the patched code changes upstream.
|
||||||
"""
|
"""
|
||||||
assert check_evaluation_loop_is_patchable()
|
assert check_evaluation_loop_is_patchable()
|
||||||
|
assert check_evaluation_loop_is_fsdp2_patchable()
|
||||||
assert check_maybe_log_save_evaluate_is_patchable()
|
assert check_maybe_log_save_evaluate_is_patchable()
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user