File size: 2,820 Bytes
e426502 22167d1 e426502 7a45936 b5f0b2d 5f33137 7a45936 29ba86d be5e88c 70199f7 47e7214 70199f7 47e7214 70199f7 47e7214 70199f7 47e7214 70199f7 47e7214 70199f7 47e7214 70199f7 47e7214 70199f7 47e7214 70199f7 47e7214 70199f7 833daf0 70199f7 29ba86d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 |
---
license: mit
language:
- en
base_model:
- ResembleAI/chatterbox
pipeline_tag: text-to-speech
tags:
- gguf-connector
---
## gguf quantized version of chatterbox
- base model from [resembleai](https://huggingface.co/ResembleAI)
- text-to-speech synthesis
### **run it with gguf-connector**
```
ggc c2
```

| Prompt | Audio Sample |
|--------|---------------|
|`Hey Connector, why your appearance looks so stupid?`<br/>`Oh, really? maybe I ate too much smart beans.`<br/>`Wow. Amazing.`<br/>`Let's go to get some more smart beans and you will become stupid as well.`<br/> | 🎧 **audio-sample-1**<br><audio controls src="https://huggingface.co/calcuis/chatterbox-gguf/resolve/main/samples%5Caudio1.wav"></audio> |
|`Now let's make my mum's favourite. So three mars bars into the pan. Then we add the tuna and just stir for a bit, just let the chocolate and fish infuse. `<br/>`A sprinkle of olive oil and some tomato ketchup. Now smell that. Oh boy this is going to be incredible.`<br/> | 🎧 **audio-sample-2**<br><audio controls src="https://huggingface.co/calcuis/chatterbox-gguf/resolve/main/samples%5Caudio2.wav"></audio> |
### **review/reference**
- simply execute the command (`ggc c2`) above in console/terminal
- opt a `vae`, a `clip(encoder)` and a `model` file in the current directory to interact with (see example below)
>
>GGUF file(s) available. Select which one for **ve**:
>
>1. s3gen-bf16.gguf
>2. s3gen-f16.gguf
>3. s3gen-f32.gguf
>4. t3_cfg-q2_k.gguf
>5. t3_cfg-q4_k_m.gguf
>6. t3_cfg-q6_k.gguf
>7. ve_fp32-f16.gguf (recommended)
>8. ve_fp32-f32.gguf
>
>Enter your choice (1 to 8): 7
>
>ve file: ve_fp32-f16.gguf is selected!
>
>GGUF file(s) available. Select which one for **t3**:
>
>1. s3gen-bf16.gguf
>2. s3gen-f16.gguf
>3. s3gen-f32.gguf
>4. t3_cfg-q2_k.gguf
>5. t3_cfg-q4_k_m.gguf (recommended)
>6. t3_cfg-q6_k.gguf
>7. ve_fp32-f16.gguf
>8. ve_fp32-f32.gguf
>
>Enter your choice (1 to 8): 5
>
>t3 file: t3_cfg-q4_k_m.gguf is selected!
>
>GGUF file(s) available. Select which one for **s3gen**:
>
>1. s3gen-bf16.gguf (recommended)
>2. s3gen-f16.gguf (for non-cuda user)
>3. s3gen-f32.gguf
>4. t3_cfg-q2_k.gguf
>5. t3_cfg-q4_k_m.gguf
>6. t3_cfg-q6_k.gguf
>7. ve_fp32-f16.gguf
>8. ve_fp32-f32.gguf
>
>Enter your choice (1 to 8): _
>
- note: for the latest update, only tokenizer will be pulled to cache automatically during the first launch; you need to prepare the **model**, **encoder** and **vae** files yourself, working like [vision](https://huggingface.co/calcuis/llava-gguf) connector right away; mix and match, more flexible
- run it entirely offline; i.e., from local URL: http://127.0.0.1:7860 with lazy webui
- gguf-connector ([pypi](https://pypi.org/project/gguf-connector)) |