pytorch
a70505cd - ns for fx: support comparing fp32 vs fp32_prepared, except shadowed (#61129)

Commit
4 years ago
ns for fx: support comparing fp32 vs fp32_prepared, except shadowed (#61129) Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/61129 Adds support the comparing fp32 model (without quantization) to a fp32 model prepared with quantization. The main missing feature was handling conv-bn fusion, since this fusion for PTQ happens outside of quantization patterns. Adds testing for this case for comparing weights and comparing activations Adds a TODO for also handling this for shadow activations, we need to first stop removing observers in graph passes before we can add this support, will be in a future PR. Test Plan: ``` python test/test_quantization.py TestFXGraphMatcherModels.test_mobilenet_v2 python test/test_quantization.py TestFXGraphMatcherModels.test_mobilenet_v2_qat python test/test_quantization.py TestFXNumericSuiteCoreAPIsModels.test_compare_activations_conv ``` Imported from OSS Reviewed By: raghuramank100 Differential Revision: D29520009 fbshipit-source-id: f63484a998f1424bd9cacf5d823b82b2edfea1ae
Author
Parents
Loading