--- pipeline_tag: text-generation tags: - llama - ggml --- **Quantization from:** [TokenBender/llama2-7b-chat-hf-codeCherryPop-qLoRA-merged](https://huggingface.co/TokenBender/llama2-7b-chat-hf-codeCherryPop-qLoRA-merged) **Converted to the GGML format with:** [llama.cpp master-b5fe67f (JUL 22, 2023)](https://github.com/ggerganov/llama.cpp/releases/tag/master-b5fe67f) **Tested with:** [koboldcpp 1.36](https://github.com/LostRuins/koboldcpp/releases/tag/v1.36) **Example usage:** ``` koboldcpp.exe llama2-7b-chat-hf-codeCherryPop-qLoRA-merged-ggmlv3.Q6_K.bin --threads 6 --contextsize 4096 --stream --smartcontext --unbantokens --ropeconfig 1.0 10000 --noblas ``` **Tested with the following format (refer to the original model and [Stanford Alpaca](https://github.com/tatsu-lab/stanford_alpaca) for additional details):** ``` ### Instruction: {code request} ### Response: ```