Hackable and optimized Transformers building blocks, supporting a composable construction.
Why do you think that https://github.com/Dao-AILab/flash-attention is a good alternative to xformers
Hackable and optimized Transformers building blocks, supporting a composable construction.
Why do you think that https://github.com/Dao-AILab/flash-attention is a good alternative to xformers