forked from Dao-AILab/flash-attention
-
Notifications
You must be signed in to change notification settings - Fork 45
Pull requests: ROCm/flash-attention
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Added Support for Rotary Positional Embeddings (both non-fused and fused kernel)
#99
opened Nov 13, 2024 by
alexkranias-amd
Loading…
Integrated Rotary Positional Embeddings (RoPEs) into flash_attn_kvcache
#83
opened Sep 27, 2024 by
alexkranias-amd
Loading…
GPUAI-1250 - Flash Attention v2.04 two modules layer_norm cannot be used fixed
#52
opened Apr 3, 2024 by
xiaoxiangAMD
Loading…
GPUAI-1250 - Flash Attention v2.04 module rotary cannot be used code fixed
#47
opened Mar 1, 2024 by
xiaoxiangAMD
Loading…
ProTip!
What’s not been updated in a month: updated:<2024-10-15.