transformers
e686fed6 - [Feature] Support using FlashAttention2 on Ascend NPU (#36696)

Commit
260 days ago
[Feature] Support using FlashAttention2 on Ascend NPU (#36696) * [Feature] Support using flash-attention on Ascend NPU * Fix qwen3 and qwen3_moe moduler conversion mismatch
Author
Parents
Loading