From da342f461120795af9c67b76431fb35e5f070136 Mon Sep 17 00:00:00 2001 From: CodemodService Bot Date: Fri, 21 Feb 2025 08:01:19 -0800 Subject: [PATCH] Auto-fix lint violations from Fixit] fbcode//pytorch/ao (#1752) Summary: Pull Request resolved: https://github.com/pytorch/ao/pull/1752 Reviewed By: amyreese Differential Revision: D69041228 --- torchao/quantization/GPTQ.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/torchao/quantization/GPTQ.py b/torchao/quantization/GPTQ.py index cb7c8d0481..b278e22b3b 100644 --- a/torchao/quantization/GPTQ.py +++ b/torchao/quantization/GPTQ.py @@ -759,7 +759,7 @@ def _create_quantized_state_dict( if self.padding_allowed: import torch.nn.functional as F - logging.warn( + logging.warning( f"warning: {fqn} is padded to satisfy in_features % 1024 == 0" ) padded_in_features = find_multiple(in_features, 1024) @@ -767,7 +767,7 @@ def _create_quantized_state_dict( weight, pad=(0, padded_in_features - in_features) ) else: - logging.warn( + logging.warning( f"warning: {fqn} is skipped, int4 requires that in_features is 32, 64, or is divisible by 1024, " + "and that groupsize and inner_k_tiles*16 evenly divide into it" ) @@ -1147,7 +1147,7 @@ def _create_quantized_state_dict( if self.padding_allowed: import torch.nn.functional as F - logging.warn( + logging.warning( f"warning: {fqn} is padded to satisfy in_features % 1024 == 0" ) padded_in_features = find_multiple(in_features, 1024) @@ -1155,7 +1155,7 @@ def _create_quantized_state_dict( weight, pad=(0, padded_in_features - in_features) ) else: - logging.warn( + logging.warning( f"warning: {fqn} is skipped, int4 requires that in_features is 32, 64, or is divisible by 1024, " + "and that groupsize and inner_k_tiles*16 evenly divide into it" )