How was this made? (question about context size)
#1
by
robbiemu
- opened
If it was llama.cpp -- did you modify the config.json to add the Qwen2ForCausalLM yarn parameters to support the long context (or would it be 32k context when run in ollama/llama.cpp)?
made with llama.cpp no mod