quark.torch.quantization.graph.processor.processor#

Module Contents#

Functions#

quark.torch.quantization.graph.processor.processor.transform_for_annotation(model: torch.fx.GraphModule) torch.fx.GraphModule#

Prepare before annotation, for both PTQ and QAT

quark.torch.quantization.graph.processor.processor.freeze_model(model: torch.fx.GraphModule) torch.fx.GraphModule#

After quantization, we need to export model (e.g onnx, torch.export), we regard the users will not need further calibration, training, optimization.