forked from Dao-AILab/flash-attention
-
Notifications
You must be signed in to change notification settings - Fork 44
Pull requests: ROCm/flash-attention
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Integrated Rotary Positional Embeddings (RoPEs) into flash_attn_kvcache
#83
opened Sep 27, 2024 by
alexkranias-amd
Loading…
GPUAI-1250 - Flash Attention v2.04 two modules layer_norm cannot be used fixed
#52
opened Apr 3, 2024 by
xiaoxiangAMD
Loading…
GPUAI-1250 - Flash Attention v2.04 module rotary cannot be used code fixed
#47
opened Mar 1, 2024 by
xiaoxiangAMD
Loading…
ProTip!
Updated in the last three days: updated:>2024-10-21.