xla
Fix a bug in flash attention where kv_seq_len should divide block_k_major.
#8671
Merged

Fix a bug in flash attention where kv_seq_len should divide block_k_major. #8671

zpcore merged 8 commits into pytorch:master from zhangp365:master
zhangp365
zhangp365 Fix a bug in flash attention where kv_seq_len should divide block_k_m…
45f6e963
qihqi qihqi requested a review from tengyifei tengyifei 270 days ago
qihqi qihqi requested a review from zpcore zpcore 270 days ago
qihqi
qihqi qihqi requested a review from qihqi qihqi 270 days ago
zpcore
zpcore commented on 2025-02-06
zpcore
zpcore commented on 2025-02-06
zpcore
zpcore commented on 2025-02-06
zpcore
zhangp365
zhangp365 add test code for flash attention and adjust previous code.
d4c84198
zhangp365 fixed a bug in test code.
cb583e66
zhangp365
zhangp365 adjust the code format in test_pallas.py and test_pallas_spmd.py by y…
5d583a6e
zhangp365 adjust the code format in test_pallas.py and test_pallas_spmd.py by y…
0a639e87
zpcore
zpcore commented on 2025-02-07
zhangp365 fix a padding bug in custom_kernel.py
cb2157f0
zhangp365 fixed a padding softmax bug in custom_kernel.py
41ac7acc
zhangp365 move the padding position in custom_kernel.py
4f0532dd
zhangp365
zhangp365
zpcore
zpcore
zpcore approved these changes on 2025-02-10
zpcore zpcore merged cff9f4e0 into master 265 days ago
zhangp365

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone