Flash Attention 2

#1
by Modularity - opened

The model does not have the support for Flash Attention 2.0, which makes it not practical

OpenGVLab org

Thanks for your feedback, flash attention is now supported.

czczup changed discussion status to closed

Is it possible to switch to Flash Attention 1.0 or disable it all together?

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment