File size: 2,820 Bytes
e426502
 
 
 
 
 
 
22167d1
 
e426502
 
 
 
7a45936
b5f0b2d
 
 
 
 
5f33137
7a45936
 
 
 
 
29ba86d
 
 
be5e88c
70199f7
 
47e7214
70199f7
47e7214
 
 
 
 
 
 
 
70199f7
47e7214
70199f7
 
 
47e7214
70199f7
47e7214
 
 
 
 
 
 
 
70199f7
47e7214
70199f7
 
 
47e7214
70199f7
47e7214
 
 
 
 
 
 
 
70199f7
47e7214
70199f7
 
833daf0
70199f7
29ba86d
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
---
license: mit
language:
- en
base_model:
- ResembleAI/chatterbox
pipeline_tag: text-to-speech
tags:
- gguf-connector
---
## gguf quantized version of chatterbox
- base model from [resembleai](https://huggingface.co/ResembleAI)
- text-to-speech synthesis

### **run it with gguf-connector**
```
ggc c2
```

![screenshot](https://raw.githubusercontent.com/calcuis/text-to-speech-synthesis-lite/master/demo.png)

| Prompt | Audio Sample |
|--------|---------------|
|`Hey Connector, why your appearance looks so stupid?`<br/>`Oh, really? maybe I ate too much smart beans.`<br/>`Wow. Amazing.`<br/>`Let's go to get some more smart beans and you will become stupid as well.`<br/> | 🎧 **audio-sample-1**<br><audio controls src="https://huggingface.co/calcuis/chatterbox-gguf/resolve/main/samples%5Caudio1.wav"></audio> |
|`Now let's make my mum's favourite. So three mars bars into the pan. Then we add the tuna and just stir for a bit, just let the chocolate and fish infuse. `<br/>`A sprinkle of olive oil and some tomato ketchup. Now smell that. Oh boy this is going to be incredible.`<br/> | 🎧 **audio-sample-2**<br><audio controls src="https://huggingface.co/calcuis/chatterbox-gguf/resolve/main/samples%5Caudio2.wav"></audio> |

### **review/reference**
- simply execute the command (`ggc c2`) above in console/terminal
- opt a `vae`, a `clip(encoder)` and a `model` file in the current directory to interact with (see example below)

>
>GGUF file(s) available. Select which one for **ve**:
>
>1. s3gen-bf16.gguf
>2. s3gen-f16.gguf
>3. s3gen-f32.gguf
>4. t3_cfg-q2_k.gguf
>5. t3_cfg-q4_k_m.gguf
>6. t3_cfg-q6_k.gguf
>7. ve_fp32-f16.gguf (recommended)
>8. ve_fp32-f32.gguf
>
>Enter your choice (1 to 8): 7
>
>ve file: ve_fp32-f16.gguf is selected!
>
>GGUF file(s) available. Select which one for **t3**:
>
>1. s3gen-bf16.gguf
>2. s3gen-f16.gguf
>3. s3gen-f32.gguf
>4. t3_cfg-q2_k.gguf
>5. t3_cfg-q4_k_m.gguf (recommended)
>6. t3_cfg-q6_k.gguf
>7. ve_fp32-f16.gguf
>8. ve_fp32-f32.gguf
>
>Enter your choice (1 to 8): 5
>
>t3 file: t3_cfg-q4_k_m.gguf is selected!
>
>GGUF file(s) available. Select which one for **s3gen**:
>
>1. s3gen-bf16.gguf (recommended)
>2. s3gen-f16.gguf (for non-cuda user)
>3. s3gen-f32.gguf
>4. t3_cfg-q2_k.gguf
>5. t3_cfg-q4_k_m.gguf
>6. t3_cfg-q6_k.gguf
>7. ve_fp32-f16.gguf
>8. ve_fp32-f32.gguf
>
>Enter your choice (1 to 8): _
>

- note: for the latest update, only tokenizer will be pulled to cache automatically during the first launch; you need to prepare the **model**, **encoder** and **vae** files yourself, working like [vision](https://huggingface.co/calcuis/llava-gguf) connector right away; mix and match, more flexible
- run it entirely offline; i.e., from local URL:  http://127.0.0.1:7860 with lazy webui
- gguf-connector ([pypi](https://pypi.org/project/gguf-connector))