transformers
[performance_optim] reduce frequency of declaring attention_mask in Ascend NPU flash attention
#38278
Merged

Loading