We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent f3fc5e7 commit 204cd48Copy full SHA for 204cd48
torchao/quantization/qat/README.md
@@ -2,7 +2,7 @@
2
3
Quantization-Aware Training (QAT) refers to applying fake quantization during the
4
training or fine-tuning process, such that the final quantized model will exhibit
5
-higher accuracies and perplexities. Fake quantization refers to rounding the float
+higher accuracies and lower perplexities. Fake quantization refers to rounding the float
6
values to quantized values without actually casting them to dtypes with lower
7
bit-widths, in contrast to post-training quantization (PTQ), which does cast the
8
quantized values to lower bit-width dtypes, e.g.:
0 commit comments