Skip to content

Commit 5b40e78

Browse files
authored
Merge branch 'main' into qbits_awq
2 parents d621f91 + 6ac7a2b commit 5b40e78

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

auto_round/auto_quantizer.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -405,8 +405,8 @@ def convert_model(self, model: nn.Module):
405405
logger.info(f"Change the dtype to `bfloat16` as {target_device.upper()} does not support float16")
406406
model = model.to(torch.bfloat16)
407407
elif "cuda" == target_device and model.dtype != torch.float16:
408-
logger.info(f"Change the dtype to `float16` for better performance")
409-
model = model.to(torch.float16)
408+
logger.info(f"Change the dtype to `float16` for better performance")
409+
model = model.to(torch.float16)
410410

411411
bits = quantization_config.bits
412412
group_size = quantization_config.group_size

0 commit comments

Comments
 (0)