diff --git a/src/llmcompressor/modifiers/awq/base.py b/src/llmcompressor/modifiers/awq/base.py index e5e8f1006..fba5f699a 100644 --- a/src/llmcompressor/modifiers/awq/base.py +++ b/src/llmcompressor/modifiers/awq/base.py @@ -1,11 +1,11 @@ import inspect from dataclasses import dataclass -from typing import Any, Callable, Dict, List, Optional, Tuple, Union +from typing import Any, Dict, List, Optional, Union import torch from compressed_tensors.utils import align_module_device, update_offload_parameter from loguru import logger -from pydantic import ConfigDict, Field +from pydantic import ConfigDict from torch.nn import Module from tqdm import tqdm @@ -16,7 +16,6 @@ from llmcompressor.utils.fsdp.helpers import get_fsdp_parent from llmcompressor.utils.helpers import calibration_forward_context from llmcompressor.utils.pytorch.module import ( - get_layer, get_layers, get_matching_layer, get_parent_by_name, @@ -146,7 +145,7 @@ class AWQModifier(Modifier): duo_scaling: bool = True _resolved_mappings: List[ResolvedMapping] = [] - _scales: Dict[str, torch.Tensor | List[torch.Tensor]] = {} + _scales: Dict[str, Union[torch.Tensor, List[torch.Tensor]]] = {} _module_kwargs: Dict = {} def on_initialize(self, state: State, **kwargs) -> bool: diff --git a/src/llmcompressor/transformers/finetune/data/__init__.py b/src/llmcompressor/transformers/finetune/data/__init__.py index 186a076bd..a53caed1b 100644 --- a/src/llmcompressor/transformers/finetune/data/__init__.py +++ b/src/llmcompressor/transformers/finetune/data/__init__.py @@ -8,7 +8,6 @@ from .flickr_30k import Flickr30K from .gsm8k import GSM8KDataset from .open_platypus import OpenPlatypusDataset -from .pile import PileValDataset from .ptb import PtbDataset from .ultrachat_200k import UltraChatDataset from .wikitext import WikiTextDataset diff --git a/src/llmcompressor/transformers/finetune/data/pile.py b/src/llmcompressor/transformers/finetune/data/pile.py deleted file mode 100644 index f420ba2a5..000000000 --- a/src/llmcompressor/transformers/finetune/data/pile.py +++ /dev/null @@ -1,27 +0,0 @@ -from copy import deepcopy -from typing import TYPE_CHECKING - -from llmcompressor.transformers.finetune.data import TextGenerationDataset -from llmcompressor.typing import Processor - -if TYPE_CHECKING: - from llmcompressor.args import DatasetArguments - - -@TextGenerationDataset.register(name="mit-han-lab/pile-val-backup", alias="pile_val") -class PileValDataset(TextGenerationDataset): - """ - Child text generation class for "The Pile" dataset - :param data_args: configuration settings for dataset loading - :param split: split from dataset to load, for instance `test` or `train[:5%]` - :param tokenizer: tokenizer to use on dataset - """ - - def __init__(self, data_args: "DatasetArguments", split: str, processor: Processor): - data_args = deepcopy(data_args) - data_args.text_column = "text" - data_args.dataset = "mit-han-lab/pile-val-backup" - super().__init__(data_args=data_args, split=split, processor=processor) - - def dataset_template(self, sample): - return {"text": sample["text"].strip()} diff --git a/tests/llmcompressor/transformers/finetune/data/test_registry.py b/tests/llmcompressor/transformers/finetune/data/test_registry.py index ce872fba9..29895b4a4 100644 --- a/tests/llmcompressor/transformers/finetune/data/test_registry.py +++ b/tests/llmcompressor/transformers/finetune/data/test_registry.py @@ -4,7 +4,6 @@ from llmcompressor.transformers.finetune.data import ( C4Dataset, OpenPlatypusDataset, - PileEvalDataset, TextGenerationDataset, WikiTextDataset, )