Flash Attention 2
#1
by
Modularity
- opened
The model does not have the support for Flash Attention 2.0, which makes it not practical
Thanks for your feedback, flash attention is now supported.
czczup
changed discussion status to
closed
Is it possible to switch to Flash Attention 1.0 or disable it all together?