Remove unused imports

This commit is contained in:
Casper
2023-12-07 21:57:51 +01:00
parent 1cb7977026
commit d9d97e3896
2 changed files with 0 additions and 11 deletions

View File

@@ -3,7 +3,6 @@
# copied from https://github.com/lm-sys/FastChat/blob/main/fastchat/train/llama_flash_attn_monkey_patch.py
import logging
from functools import partial
from typing import List, Optional, Tuple, Union
import torch

View File

@@ -6,17 +6,7 @@ from typing import List, Optional, Tuple, Union
import torch
import transformers
from einops import rearrange
from flash_attn.bert_padding import pad_input, unpad_input
from flash_attn.flash_attn_interface import ( # pylint: disable=ungrouped-imports
flash_attn_kvpacked_func,
flash_attn_varlen_kvpacked_func,
flash_attn_varlen_qkvpacked_func,
)
from transformers.modeling_outputs import BaseModelOutputWithPast
from transformers.models.mistral.modeling_mistral import (
MistralAttention as OriginalMistralAttention,
)
from transformers.models.mistral.modeling_mistral import (
MistralDecoderLayer as OriginalMistralDecoderLayer,
MistralMLP