pytorch
2c748b75 - [ONNX] Trace model if quantization is detected

Commit
2 years ago
[ONNX] Trace model if quantization is detected Previously pre-tracing model is required for exporting quantized model. e.g. calling `traced_m = torch.jit.trace(model, inputs)` and export `traced_m`. The reason was quantized weights are stored in a unique `PackedParam` structure, and they need to be handled by tracing to be exportable. This PR enables export api to call tracing underneath if it detects quantization in the model. Pull Request resolved: https://github.com/pytorch/pytorch/pull/75921 Approved by: https://github.com/garymm
Author
Committer
Parents
Loading