On this page
prepare_qat
class torch.quantization.prepare_qat(model, mapping=None, inplace=False)[source]-
Prepares a copy of the model for quantization calibration or quantization-aware training and converts it to quantized version.
Quantization configuration should be assigned preemptively to individual submodules in
.qconfigattribute.- Parameters:
-
- model – input model to be modified in-place
- mapping – dictionary that maps float modules to quantized modules to be replaced.
- inplace – carry out model transformations in-place, the original module is mutated
© 2024, PyTorch Contributors
PyTorch has a BSD-style license, as found in the LICENSE file.
https://pytorch.org/docs/1.13/generated/torch.quantization.prepare_qat.html