[ONNX] Add quantization support to more single output ops (#83008)
#80039
- Implement quantization support for single output ops
- quantized::sigmoid
- quantized::instance_norm
- aten::reshape
- aten::reshape_as
- aten::sum
- aten::mean
- aten::prod
- aten::t
- aten::numpy_T
- aten::expand
- aten::expand_as
- aten::embedding
- aten::embedding_bag
- aten::view
- aten::select
- aten::eq
- aten::ne
- aten::gt
- aten::lt
- aten::le
- aten::ge
- quantized::layer_norm
- aten::elu
- aten::selu
- aten::maximum
- aten::minimum
- aten::amax
- aten::amin
- aten::hardtanh
- aten::hardswish
- quantized::group_norm
- aten::as_strided
- quantized::leaky_relu
- aten::transpose
- Avoid modifying functions in `quantized_args` and have the wrapper closed over `scale` and `zero_point` instead (for purity)
- Remove magic number and assign it to INT64_MAX
- implement `_unpack_quantized_tensor` for handling quantized tensor unpacking to separate the logic from tuple unpacking and for clearer error handling
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83008
Approved by: https://github.com/BowenBao