transformers
f834ca2c - Attention Quantization with FBGemm & TP (#37384)

Commit
252 days ago
Attention Quantization with FBGemm & TP (#37384) * fix * keep fused * contiguous * rm print * update * update * rm print
Author
Parents
Loading