Closed
Description
https://github.com/Dao-AILab/flash-attention
Flash attention v2 was released claiming 2x speedups. Making an issue to remind myself to have a look at it. And also if anyone else wants to try implement it.
https://github.com/Dao-AILab/flash-attention
Flash attention v2 was released claiming 2x speedups. Making an issue to remind myself to have a look at it. And also if anyone else wants to try implement it.