transformers
28c9541c
- Attention Quantization with FBGemm & TP (#37384)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
282 days ago
Attention Quantization with FBGemm & TP (#37384) * fix * keep fused * contiguous * rm print * update * update * rm print
Author
MekkCyber
Committer
ArthurZucker
Parents
56512077
Loading