[CUDA] Update Flash Attention Implementation and APIs #26937
Update Flash Attention Implementation and APIs
50dbc150
tianleiwu
changed the title Update Flash Attention Implementation and APIs [CUDA] Update Flash Attention Implementation and APIs 56 days ago
update base_dir
4f0942fa
cleanup of redundant default assignments
ee4205f3
Merge 'origin/main' into tlwu/refresh_flash_attention
0b896afd
tianleiwu
dismissed their stale review
via 0b896afd
54 days ago
tianleiwu
enabled auto-merge (squash) 54 days ago
tianleiwu
merged
5d24c954
into main 53 days ago
tianleiwu
deleted the tlwu/refresh_flash_attention branch 53 days ago
Assignees
No one assigned
Login to write a write a comment.
Login via GitHub