onnxruntime
46caf476 - [QNN EP] Fix 16x16 MatMul translation (#24846)

Commit
211 days ago
[QNN EP] Fix 16x16 MatMul translation (#24846) ### Description - QNN's 16x16 FC doesn't support asymmetric int16 weight - QNN's 16x16 MatMul doesn't support asymmetric int16 weight initializer. - Insert Convert Op to convert from asymmetric uint16 weight to symmetric int16 weight. - Add unit tests to verify 16x16 MatMul translations. ### Motivation and Context - This fix schedules 16x16 MatMul Ops on QNN HTP accelerator. - This improves inference time of Models contain 16x16 MatMul operators
Author
Parents
Loading