Quantise to support llama.cpp

#1
by TusharRay - opened

Can this be quantised to support https://github.com/ggerganov/llama.cpp ? The llama.cpp is really performant and this model can then be widely used across multiple platforms!

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment