transformers
Proper performant flex attention implementation
#36103
Closed

Proper performant flex attention implementation #36103

bursteratom wants to merge 22 commits into huggingface:main from bursteratom:proper_flex
bursteratom
vasqu
vasqu commented on 2025-02-09
vasqu
ArthurZucker
ArthurZucker commented on 2025-02-10
molbap
molbap commented on 2025-02-10
bursteratom bursteratom force pushed from 516de450 to ae9a2b02 205 days ago
bursteratom bursteratom requested a review from molbap molbap 205 days ago
bursteratom bursteratom requested a review from vasqu vasqu 205 days ago
bursteratom bursteratom requested a review from ArthurZucker ArthurZucker 205 days ago
bursteratom
bursteratom bursteratom force pushed from 4c525764 to 7519b3cc 204 days ago
bursteratom bursteratom force pushed from 7519b3cc to 8c28c9db 204 days ago
vasqu
vasqu approved these changes on 2025-02-12
bursteratom bursteratom force pushed from 8e50735f to e705da7b 203 days ago
bursteratom bursteratom force pushed from 432bafaa to 74833140 203 days ago
ArthurZucker
ArthurZucker commented on 2025-02-14
bursteratom bursteratom force pushed from 5a0cb2dd to 1dea5a8f 199 days ago
bursteratom bursteratom force pushed from fb9c4c63 to 3d9377fb 195 days ago
bursteratom bursteratom force pushed from 0bba7283 to 0c200a05 191 days ago
bursteratom bursteratom force pushed from 2406dabc to ff2a4556 191 days ago
bursteratom bursteratom force pushed from 99e62c0a to c50468c1 191 days ago
bursteratom bursteratom force pushed from 8aaeda8c to 864efb28 187 days ago
ArthurZucker
ArthurZucker approved these changes on 2025-03-01
shethaadit
shethaadit approved these changes on 2025-03-03
bursteratom proper performant flex attention implementation
800a7e70
bursteratom wrapper for flex attention to compile only when triggered
c331bb3f
bursteratom wrapper for flex attention to compile only when triggered
e1438ade
bursteratom attention mask type detection
68bd4e6a
bursteratom Update src/transformers/integrations/flex_attention.py
cf0ad129
bursteratom nit
2afa102d
bursteratom nit
a78f7bc1
bursteratom nit
9d1ee83d
bursteratom nit
c2691909
bursteratom gemma2 support
4e58c63f
bursteratom add citation for torchtune
6237ae4a
bursteratom Update src/transformers/models/llama/modeling_llama.py
eb254a8d
bursteratom Update flex_attention.py
f593d3a8
bursteratom nit
6cf7ea91
bursteratom nit
4da79476
bursteratom nit
743ab13c
bursteratom reset gemma2 modifications
5ab35829
bursteratom nit
ad63890e
bursteratom nit
f3b3bae7
bursteratom nit
43871f7e
bursteratom licencing
f0945806
bursteratom bursteratom force pushed from 864efb28 to f0945806 184 days ago
bursteratom
ArthurZucker
ArthurZucker ArthurZucker added flex attention
ArthurZucker ArthurZucker added Compilation
ArthurZucker
ArthurZucker
bursteratom Merge branch 'main' into proper_flex
b81e90f1
bursteratom bursteratom closed this 177 days ago
bursteratom bursteratom reopened this 177 days ago
github-actions
github-actions github-actions marked this pull request as draft 177 days ago
bursteratom bursteratom closed this 177 days ago

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone