transformers
f834ca2c
- Attention Quantization with FBGemm & TP (#37384)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
252 days ago
Attention Quantization with FBGemm & TP (#37384) * fix * keep fused * contiguous * rm print * update * update * rm print
References
#37384 - Attention Quantization with FBGemm & TP
Author
MekkCyber
Parents
c5c648dd
Loading