transformers
afa79da8
- no need to pad tesnors to 16 byte strides if we made sure our tiny testing models have 16 byte aligned weights
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
6 days ago
no need to pad tesnors to 16 byte strides if we made sure our tiny testing models have 16 byte aligned weights
References
#42697 - batched and grouped experts implementations
Author
IlyasMoutawwakil
Parents
c9745f3f
Loading