Compare commits

..

16 Commits

Author SHA1 Message Date
Salman Mohammadi
7d8e8c9ac2 nit [skip-e2e] 2025-08-13 12:58:30 +01:00
Salman Mohammadi
7c2466b739 nit 2025-08-13 12:58:13 +01:00
Salman Mohammadi
3146cb56dd docs 2025-08-13 12:53:58 +01:00
Salman Mohammadi
c09b0a3bbf reverting change 2025-08-13 11:27:15 +01:00
Salman Mohammadi
e05acccd77 linting 2025-08-13 11:24:22 +01:00
Salman Mohammadi
c44abad531 debugging CI 2025-08-13 11:24:05 +01:00
Salman Mohammadi
817d70e669 debugging CI 2025-08-13 10:45:41 +01:00
Salman Mohammadi
03f5a7fd16 adding back 2025-08-12 18:35:12 +01:00
Salman Mohammadi
3d9b96a94f testing revert 2025-08-12 15:53:43 +01:00
Salman Mohammadi
42c16024a2 docs 2025-08-12 15:34:46 +01:00
Salman Mohammadi
ec94d632f3 docs 2025-08-12 14:07:55 +01:00
Salman Mohammadi
e8bd3b0b3b Merge branch 'fix-preview' of github.com:axolotl-ai-cloud/axolotl into fix-preview 2025-08-12 13:42:56 +01:00
Salman Mohammadi
5a08b94668 update workflow 2025-08-12 12:29:09 +01:00
salman
ecb8c1f4b3 Merge branch 'main' into fix-preview 2025-08-12 09:43:39 +01:00
Salman Mohammadi
ab57be6526 render docs on python file change to preview api ref 2025-08-12 09:43:23 +01:00
Wing Lian
c9640bca2c attempt to fix quartodoc render for yields 2025-08-10 22:23:09 -04:00
6 changed files with 15 additions and 16 deletions

View File

@@ -57,6 +57,13 @@ We welcome ideas for improvements and new features. To suggest an enhancement, o
5. Push your branch to your fork on GitHub.
6. Open a new pull request against the `main` branch of the axolotl repository. Include a clear and concise description of your changes, referencing any related issues.
#### Skipping CI Checks
You can skip certain CI checks by including specific keywords in your commit messages:
- `[skip ci]` or `skip ci` - Skips all CI checks for that commit
- `[skip-e2e]` or `skip-e2e` - Skips only end-to-end tests while running other CI checks
## Style Guidelines
### Code Style

10
TODO.md
View File

@@ -1,10 +0,0 @@
# todo list
- [] Validation of parameters for combinations that won't work
## things that are known not to work
- FSDP offload and gradient_checkpointing - https://github.com/pytorch/pytorch/issues/82203
- adamw_bnb_8bit doesn't play well with FSDP offload

View File

@@ -14,7 +14,7 @@ packaging==23.2
huggingface_hub>=0.33.0
peft==0.17.0
transformers @ git+https://github.com/vasqu/transformers@fix-fa-integration
transformers==4.55.0
tokenizers>=0.21.1
accelerate==1.10.0
datasets==4.0.0

View File

@@ -67,14 +67,12 @@ def build_command(base_cmd: list[str], options: dict[str, Any]) -> list[str]:
def generate_config_files(config: str, sweep: str | None) -> Iterator[tuple[str, bool]]:
"""
Generate list of configuration files to process.
Generate list of configuration files to process. Yields a tuple of the configuration file name and a boolean indicating
whether this is a group of configurations (i.e., a sweep).
Args:
config: Base configuration file
sweep: Sweep configuration file
Yields:
Tuple of configuration file name and whether this is a group of configurations
"""
if not sweep:

View File

@@ -18,7 +18,9 @@ from torch.distributed import DeviceMesh
try:
from transformers.modeling_flash_attention_utils import _flash_supports_window
except ImportError:
_flash_supports_window = True
from transformers.modeling_flash_attention_utils import (
_flash_supports_window_size as _flash_supports_window,
)
from axolotl.monkeypatch.utils import get_cu_seqlens_from_pos_ids
from axolotl.utils.logging import get_logger

View File

@@ -3,6 +3,7 @@
import unittest
from axolotl.monkeypatch.transformers.trainer_loss_calc import (
check_evaluation_loop_is_fsdp2_patchable,
check_evaluation_loop_is_patchable,
check_maybe_log_save_evaluate_is_patchable,
)
@@ -19,6 +20,7 @@ class TestTrainerLossCalc(unittest.TestCase):
the patched code changes upstream.
"""
assert check_evaluation_loop_is_patchable()
assert check_evaluation_loop_is_fsdp2_patchable()
assert check_maybe_log_save_evaluate_is_patchable()