From b0cce2435f0b72d8d8a6f0dc6b18dc409160b394 Mon Sep 17 00:00:00 2001 From: Ishan Padalkar <35197825+ishandeva@users.noreply.github.com> Date: Tue, 8 Oct 2024 16:38:56 +0530 Subject: [PATCH] Applied formatter changes Ran ruff formatter on the project (`check_dirs` in makefile) to fix format issues. --- examples/nlp/text-generation/quantize_causal_lm_model.py | 4 +++- examples/vision/object-detection/quantize_owl_model.py | 1 - optimum/quanto/library/extensions/extension.py | 1 - optimum/quanto/models/diffusers_models.py | 2 -- optimum/quanto/models/transformers_models.py | 2 -- optimum/quanto/subpackage/commands/base.py | 1 - optimum/quanto/tensor/activations/qbytes.py | 1 - optimum/quanto/tensor/optimizers/absmax_optimizer.py | 1 - optimum/quanto/tensor/optimizers/affine_optimizer.py | 1 - optimum/quanto/tensor/optimizers/max_optimizer.py | 1 - optimum/quanto/tensor/optimizers/optimizer.py | 1 - optimum/quanto/tensor/optimizers/symmetric_optimizer.py | 1 - optimum/quanto/tensor/qbits.py | 1 - optimum/quanto/tensor/qbytes.py | 1 - optimum/quanto/tensor/qtensor.py | 1 - optimum/quanto/tensor/weights/marlin/fp8/qbits.py | 1 - optimum/quanto/tensor/weights/qbits.py | 1 - optimum/quanto/tensor/weights/qbytes.py | 1 - test/library/test_quantize.py | 4 +--- test/quantize/test_quantize_patterns.py | 1 - test/tensor/weights/test_weight_qbits_tensor_quantize.py | 4 +--- 21 files changed, 5 insertions(+), 27 deletions(-) diff --git a/examples/nlp/text-generation/quantize_causal_lm_model.py b/examples/nlp/text-generation/quantize_causal_lm_model.py index da279ab2..1b96cea9 100644 --- a/examples/nlp/text-generation/quantize_causal_lm_model.py +++ b/examples/nlp/text-generation/quantize_causal_lm_model.py @@ -120,7 +120,9 @@ def main(): torch_dtype = ( torch.float16 if args.load_dtype == "float16" - else torch.bfloat16 if args.load_dtype == "bfloat16" else torch.float32 + else torch.bfloat16 + if args.load_dtype == "bfloat16" + else torch.float32 ) model = AutoModelForCausalLM.from_pretrained(args.model, torch_dtype=torch_dtype, low_cpu_mem_usage=True).to( device diff --git a/examples/vision/object-detection/quantize_owl_model.py b/examples/vision/object-detection/quantize_owl_model.py index 26c3c45d..92312087 100644 --- a/examples/vision/object-detection/quantize_owl_model.py +++ b/examples/vision/object-detection/quantize_owl_model.py @@ -12,7 +12,6 @@ def detect(model, processor, image, texts): - inputs = processor(text=texts, images=image, return_tensors="pt").to(model.device) # forward pass diff --git a/optimum/quanto/library/extensions/extension.py b/optimum/quanto/library/extensions/extension.py index 23ce0e6f..03257e2f 100644 --- a/optimum/quanto/library/extensions/extension.py +++ b/optimum/quanto/library/extensions/extension.py @@ -11,7 +11,6 @@ class Extension(object): - def __init__( self, name: str, diff --git a/optimum/quanto/models/diffusers_models.py b/optimum/quanto/models/diffusers_models.py index 9130b620..5914a93a 100644 --- a/optimum/quanto/models/diffusers_models.py +++ b/optimum/quanto/models/diffusers_models.py @@ -42,7 +42,6 @@ class QuantizedDiffusersModel(ModelHubMixin): - BASE_NAME = "quanto" base_class = None @@ -188,5 +187,4 @@ def _save_pretrained(self, save_directory: Path) -> None: class QuantizedPixArtTransformer2DModel(QuantizedDiffusersModel): - base_class = PixArtTransformer2DModel diff --git a/optimum/quanto/models/transformers_models.py b/optimum/quanto/models/transformers_models.py index 24580a0d..39a6dc55 100644 --- a/optimum/quanto/models/transformers_models.py +++ b/optimum/quanto/models/transformers_models.py @@ -36,7 +36,6 @@ class QuantizedTransformersModel(ModelHubMixin): - BASE_NAME = "quanto" auto_class = None @@ -178,5 +177,4 @@ def _save_pretrained(self, save_directory: Path) -> None: class QuantizedModelForCausalLM(QuantizedTransformersModel): - auto_class = AutoModelForCausalLM diff --git a/optimum/quanto/subpackage/commands/base.py b/optimum/quanto/subpackage/commands/base.py index a0dd0dff..95f25777 100644 --- a/optimum/quanto/subpackage/commands/base.py +++ b/optimum/quanto/subpackage/commands/base.py @@ -23,7 +23,6 @@ @optimum_cli_subcommand() class QuantoCommand(BaseOptimumCLICommand): - COMMAND = CommandInfo(name="quanto", help="Hugging Face models quantization tools") SUBCOMMANDS = ( CommandInfo( diff --git a/optimum/quanto/tensor/activations/qbytes.py b/optimum/quanto/tensor/activations/qbytes.py index 0b06fd1d..6efaae03 100644 --- a/optimum/quanto/tensor/activations/qbytes.py +++ b/optimum/quanto/tensor/activations/qbytes.py @@ -26,7 +26,6 @@ class ActivationQBytesQuantizer(Function): - @staticmethod def forward(ctx, base: torch.Tensor, qtype: qtype, scale: torch.Tensor) -> torch.Tensor: if qtype.bits != 8: diff --git a/optimum/quanto/tensor/optimizers/absmax_optimizer.py b/optimum/quanto/tensor/optimizers/absmax_optimizer.py index 4db7dd5b..88cd95a0 100644 --- a/optimum/quanto/tensor/optimizers/absmax_optimizer.py +++ b/optimum/quanto/tensor/optimizers/absmax_optimizer.py @@ -24,7 +24,6 @@ class AbsmaxOptimizer(SymmetricOptimizer): - def optimize( self, base: torch.Tensor, qtype: qtype, axis: Optional[int] = None ) -> Union[torch.Tensor, Tuple[torch.Tensor, torch.Tensor]]: diff --git a/optimum/quanto/tensor/optimizers/affine_optimizer.py b/optimum/quanto/tensor/optimizers/affine_optimizer.py index 71b9cb61..80cf091a 100644 --- a/optimum/quanto/tensor/optimizers/affine_optimizer.py +++ b/optimum/quanto/tensor/optimizers/affine_optimizer.py @@ -25,7 +25,6 @@ class AffineOptimizer(Optimizer): - def __call__( self, base: torch.Tensor, diff --git a/optimum/quanto/tensor/optimizers/max_optimizer.py b/optimum/quanto/tensor/optimizers/max_optimizer.py index 8b4f5f62..a300be7c 100644 --- a/optimum/quanto/tensor/optimizers/max_optimizer.py +++ b/optimum/quanto/tensor/optimizers/max_optimizer.py @@ -24,7 +24,6 @@ class MaxOptimizer(AffineOptimizer): - def optimize( self, base: torch.Tensor, qtype: qtype, axis: int ) -> Union[torch.Tensor, Tuple[torch.Tensor, torch.Tensor]]: diff --git a/optimum/quanto/tensor/optimizers/optimizer.py b/optimum/quanto/tensor/optimizers/optimizer.py index 09b8c9f9..f4a37bde 100644 --- a/optimum/quanto/tensor/optimizers/optimizer.py +++ b/optimum/quanto/tensor/optimizers/optimizer.py @@ -22,7 +22,6 @@ class Optimizer(ABC): - def __call__( self, base: torch.Tensor, bits: int, axis: int, group_size: Optional[int] = None ) -> Union[torch.Tensor, Tuple[torch.Tensor, torch.Tensor]]: diff --git a/optimum/quanto/tensor/optimizers/symmetric_optimizer.py b/optimum/quanto/tensor/optimizers/symmetric_optimizer.py index 87a817bf..fa5c3dbc 100644 --- a/optimum/quanto/tensor/optimizers/symmetric_optimizer.py +++ b/optimum/quanto/tensor/optimizers/symmetric_optimizer.py @@ -24,7 +24,6 @@ class SymmetricOptimizer(Optimizer): - def __call__(self, base: torch.Tensor, qtype: qtype, axis: Optional[int] = None) -> torch.Tensor: if axis not in [None, 0, -1]: raise ValueError("axis parameter must be None, 0 (first axis) or -1 (last axis)") diff --git a/optimum/quanto/tensor/qbits.py b/optimum/quanto/tensor/qbits.py index 17b35639..6f0b5e58 100644 --- a/optimum/quanto/tensor/qbits.py +++ b/optimum/quanto/tensor/qbits.py @@ -54,7 +54,6 @@ def backward(ctx, gO): class QBitsTensor(QTensor): - def __init__(self, qtype, axis, group_size, size, stride, data, scale, shift, requires_grad=False): super().__init__(qtype, axis) self._data = data diff --git a/optimum/quanto/tensor/qbytes.py b/optimum/quanto/tensor/qbytes.py index 79d9577e..b18d6f4d 100644 --- a/optimum/quanto/tensor/qbytes.py +++ b/optimum/quanto/tensor/qbytes.py @@ -37,7 +37,6 @@ def backward(ctx, gO): class QBytesTensor(QTensor): - def __init__(self, qtype, axis, size, stride, data, scale, requires_grad=False): super().__init__(qtype, axis) self._data = data diff --git a/optimum/quanto/tensor/qtensor.py b/optimum/quanto/tensor/qtensor.py index 2ce4c8d4..5c8ee856 100644 --- a/optimum/quanto/tensor/qtensor.py +++ b/optimum/quanto/tensor/qtensor.py @@ -30,7 +30,6 @@ def qfallback(callable, *args, **kwargs): class QTensor(torch.Tensor): - def __init__(self, qtype, axis): self._qtype = qtype self._axis = axis diff --git a/optimum/quanto/tensor/weights/marlin/fp8/qbits.py b/optimum/quanto/tensor/weights/marlin/fp8/qbits.py index 3b9db6d0..0cf24243 100644 --- a/optimum/quanto/tensor/weights/marlin/fp8/qbits.py +++ b/optimum/quanto/tensor/weights/marlin/fp8/qbits.py @@ -61,7 +61,6 @@ def __new__(cls, qtype, axis, size, stride, data, scale, requires_grad=False): ) def __init__(self, qtype, axis, size, stride, data, scale, requires_grad=False): - assert axis == 0 assert data.ndim == 2 diff --git a/optimum/quanto/tensor/weights/qbits.py b/optimum/quanto/tensor/weights/qbits.py index 3afce3f5..f9ca965c 100644 --- a/optimum/quanto/tensor/weights/qbits.py +++ b/optimum/quanto/tensor/weights/qbits.py @@ -31,7 +31,6 @@ class WeightsQBitsQuantizer(Function): - @staticmethod def forward( ctx, diff --git a/optimum/quanto/tensor/weights/qbytes.py b/optimum/quanto/tensor/weights/qbytes.py index 6d316218..68d0f65b 100644 --- a/optimum/quanto/tensor/weights/qbytes.py +++ b/optimum/quanto/tensor/weights/qbytes.py @@ -28,7 +28,6 @@ class WeightQBytesQuantizer(Function): - @staticmethod def forward( ctx, base: torch.Tensor, qtype: qtype, axis: int, scale: torch.Tensor, activation_qtype: qtype, optimized: bool diff --git a/test/library/test_quantize.py b/test/library/test_quantize.py index bfbd784f..21d8d88b 100644 --- a/test/library/test_quantize.py +++ b/test/library/test_quantize.py @@ -96,9 +96,7 @@ def test_affine_quantize(input_shape, dtype, qtype, axis, group_size, shift_mode "zeropoint": 6e-2, "float": 5e-2, }, - }[ - qtype - ][shift_mode] + }[qtype][shift_mode] if group_size is not None: qa = ungroup(qa, axis=axis, orig_shape=a.shape) assert_similar(a, qa, atol=atol) diff --git a/test/quantize/test_quantize_patterns.py b/test/quantize/test_quantize_patterns.py index 505066da..11b85d8d 100644 --- a/test/quantize/test_quantize_patterns.py +++ b/test/quantize/test_quantize_patterns.py @@ -36,7 +36,6 @@ def forward(self, inputs): class ClassificationModel(torch.nn.Module): - def __init__(self, input_size, output_size, hidden_size, classes): super().__init__() self.model = MLP(input_size, output_size, hidden_size) diff --git a/test/tensor/weights/test_weight_qbits_tensor_quantize.py b/test/tensor/weights/test_weight_qbits_tensor_quantize.py index 149c8ff4..5453993e 100644 --- a/test/tensor/weights/test_weight_qbits_tensor_quantize.py +++ b/test/tensor/weights/test_weight_qbits_tensor_quantize.py @@ -49,9 +49,7 @@ def test_weight_qbits_tensor_quantize(input_shape, dtype, qtype, axis, group_siz "zeropoint": 6e-2, "float": 5e-2, }, - }[ - qtype - ][shift_mode] + }[qtype][shift_mode] assert_similar(a, qa, atol=atol)