From 1720babdcc905c335bf9d89fec013ab1891e61af Mon Sep 17 00:00:00 2001 From: Marc Sun Date: Tue, 12 Dec 2023 20:12:12 +0100 Subject: [PATCH] fix log --- optimum/gptq/quantizer.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/optimum/gptq/quantizer.py b/optimum/gptq/quantizer.py index 0aa186425a..96071c53cb 100644 --- a/optimum/gptq/quantizer.py +++ b/optimum/gptq/quantizer.py @@ -221,7 +221,7 @@ def convert_model(self, model: nn.Module): layers_to_keep = sum(self.modules_in_block_to_quantize, []) for name in list(layers_to_be_replaced.keys()): if not any(name.endswith(layer) for layer in layers_to_keep): - logger.info(f"Quantization disabled for {name} (only modules_in_block_to_quantize={modules_in_block_to_quantize} are quantized)") + logger.info(f"Quantization disabled for {name} (only modules_in_block_to_quantize={self.modules_in_block_to_quantize} are quantized)") del layers_to_be_replaced[name] self._replace_by_quant_layers(model, layers_to_be_replaced) return model