tpmccallum's picture
Update README.md
8d06f06
---
license: cc-by-4.0
---
The [GGML model](https://huggingface.co/tpmccallum/llama-2-13b-deep-haiku-GGML/blob/main/llama-2-13b-deep-haiku.ggml.fp16.bin) contained herein was generated by following the successive process in the Collab environments at https://github.com/robgon-art/DeepHaiku-LLaMa.
> **NOTE:** This tutorial uses [Meta AI's Llama-2-13b-chat-hf](https://huggingface.co/meta-llama/Llama-2-13b-chat-hf) as the starting point (before the above quantizing process is performed). Therefore you will need to visit [Meta's Llama webpage](https://ai.meta.com/resources/models-and-libraries/llama-downloads/) and agree to Meta's License, Acceptable Use Policy, and to Meta’s privacy policy before fetching and using Llama models.
> **TIP:** The only change (aside from adding usernames and access tokens) was to substitute the following code (note the downgrading of `llama.cpp` using `checkout cf348a6` below; this is so that the Collab created the older GGML version of the model instead of the newer GGUF version) in the [2_Deep_Haiku_Quantize_Model_to_GGML](https://github.com/robgon-art/DeepHaiku-LLaMa/blob/main/2_Deep_Haiku_Quantize_Model_to_GGML.ipynb) part of the aforementioned process:
```
!rm -rf llama.cpp
!git clone https://github.com/ggerganov/llama.cpp
!cd llama.cpp && git pull && git checkout cf348a6 && make clean && LLAMA_CUBLAS=1 make
!pip install numpy==1.23
```