We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
2 parents d621f91 + 6ac7a2b commit 5b40e78Copy full SHA for 5b40e78
auto_round/auto_quantizer.py
@@ -405,8 +405,8 @@ def convert_model(self, model: nn.Module):
405
logger.info(f"Change the dtype to `bfloat16` as {target_device.upper()} does not support float16")
406
model = model.to(torch.bfloat16)
407
elif "cuda" == target_device and model.dtype != torch.float16:
408
- logger.info(f"Change the dtype to `float16` for better performance")
409
- model = model.to(torch.float16)
+ logger.info(f"Change the dtype to `float16` for better performance")
+ model = model.to(torch.float16)
410
411
bits = quantization_config.bits
412
group_size = quantization_config.group_size
0 commit comments