transformers
[performance_optim] reduce frequency of declaring attention_mask in Ascend NPU flash attention
#38278
Merged
Go
Login via GitHub
Home
Pricing
FAQ
Install
Login
via GitHub
Overview
Commits
1
Changes
View On
GitHub
Loading