pytorch
0f802eed - [Quant][FX] Lower QConvAddReLU2d for onednn backend (#91155)

Commit
1 year ago
[Quant][FX] Lower QConvAddReLU2d for onednn backend (#91155) **Summary** Add quantization mappings for QConvAddReLU2d for int8 inference for onednn backend. The fusion and lowering is supported only in FX mode. **Test plan** ``` python -m pytest test_quantization.py -k test_fuse_conv_bn_add_relu_onednn python -m pytest test_quantization.py -k test_fuse_conv_bn_add_relu_by_default python -m pytest test_quantization.py -k test_fuse_conv_bn_add_relu_lowering ``` Pull Request resolved: https://github.com/pytorch/pytorch/pull/91155 Approved by: https://github.com/jgong5, https://github.com/jerryzh168
Committer
Parents
Loading