Add flash attention for `gpt_bigcode` #26479
added flash attention of gpt_bigcode
447703e7
changed docs
7baa248a
Update src/transformers/models/gpt_bigcode/modeling_gpt_bigcode.py
bddd8e68
add FA-2 docs
7f38f86b
Merge branch 'flash_attn_starcoder' of https://github.com/susnato/tra…
50506faf
oops
28ddca39
Update docs/source/en/perf_infer_gpu_one.md Last Nit
542c2759
Ltrack
approved these changes
on 2023-10-14
Merge remote-tracking branch 'upstream/main' into HEAD
72b353bb
fix
f43ec5a4
oops
b2aa0d9b
Merge branch 'main' into flash_attn_starcoder
4792b337
remove padding_mask
22a64cbb
change getattr->hasattr logic
ba0de166
changed .md file
d577b4f9
Assignees
No one assigned
Login to write a write a comment.
Login via GitHub