view post Post 233 You can get flash-attention 3 ⚡️ directly from the hub now using kernels! kernels-community/flash-attn3 See translation
danieldk/Qwen2.5-1.5B-Instruct-w8a8-int-dynamic-weight Text Generation • 2B • Updated Nov 14, 2024 • 1.33k •