mlx-lm fails to quantize the larger sibling of the linear attention model (Ring Flash Linear v2)
· Sign up or log in to comment