fix
This commit is contained in:
@@ -18,8 +18,8 @@ from pydantic import (
|
|||||||
)
|
)
|
||||||
from transformers.utils.import_utils import is_torch_npu_available
|
from transformers.utils.import_utils import is_torch_npu_available
|
||||||
|
|
||||||
from axolotl.utils.distributed import is_main_process
|
|
||||||
from axolotl.monkeypatch.attention.ring_attn import RingAttnFunc
|
from axolotl.monkeypatch.attention.ring_attn import RingAttnFunc
|
||||||
|
from axolotl.utils.distributed import is_main_process
|
||||||
from axolotl.utils.schemas.datasets import (
|
from axolotl.utils.schemas.datasets import (
|
||||||
DatasetConfig,
|
DatasetConfig,
|
||||||
DPODataset,
|
DPODataset,
|
||||||
@@ -261,7 +261,7 @@ class AxolotlInputConfig(
|
|||||||
|
|
||||||
sequence_parallel_degree: int | None = None
|
sequence_parallel_degree: int | None = None
|
||||||
heads_k_stride: int | None = None
|
heads_k_stride: int | None = None
|
||||||
ring_attn_func: RingAttnFunc | None = None
|
ring_attn_func: str | None = None
|
||||||
|
|
||||||
special_tokens: SpecialTokensConfig | None = None
|
special_tokens: SpecialTokensConfig | None = None
|
||||||
tokens: list[str] | None = None
|
tokens: list[str] | None = None
|
||||||
|
|||||||
Reference in New Issue
Block a user