transformers
0327d0f7 - [performance_optim] define flash attention mask on NPU device directly (#37698)

Commit
334 days ago
[performance_optim] define flash attention mask on NPU device directly (#37698) Co-authored-by: Mohamed Mekkouri <93391238+MekkCyber@users.noreply.github.com>
Author
Parents
Loading