How was this made? (question about context size)

#1
by robbiemu - opened

If it was llama.cpp -- did you modify the config.json to add the Qwen2ForCausalLM yarn parameters to support the long context (or would it be 32k context when run in ollama/llama.cpp)?

made with llama.cpp no mod

Sign up or log in to comment