transformers
e686fed6
- [Feature] Support using FlashAttention2 on Ascend NPU (#36696)
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Commit
View On
GitHub
Commit
260 days ago
[Feature] Support using FlashAttention2 on Ascend NPU (#36696) * [Feature] Support using flash-attention on Ascend NPU * Fix qwen3 and qwen3_moe moduler conversion mismatch
References
#36696 - [Feature] Support using FlashAttention2 on Ascend NPU
Author
FightingZhen
Parents
a03cee7a
Loading