Spaces:
Running
Running
Rasmus Lellep
commited on
Commit
·
86125cc
1
Parent(s):
d9297af
gradio changes
Browse files- .DS_Store +0 -0
- app.py +1 -6
- app_local.py +0 -1
- requirements.txt +0 -1
.DS_Store
CHANGED
|
Binary files a/.DS_Store and b/.DS_Store differ
|
|
|
app.py
CHANGED
|
@@ -50,8 +50,6 @@ os.chmod("ffmpeg", st.st_mode | stat.S_IEXEC)
|
|
| 50 |
# This will trigger downloading model
|
| 51 |
print("Downloading Coqui XTTS V2")
|
| 52 |
|
| 53 |
-
#model_name = "tts_models/multilingual/multi-dataset/xtts_v2"
|
| 54 |
-
#ModelManager().download_model(model_name)
|
| 55 |
model_folder = "model/"
|
| 56 |
config_file = "config.json"
|
| 57 |
if not os.path.exists(model_folder + config_file):
|
|
@@ -87,7 +85,6 @@ model.load_checkpoint(
|
|
| 87 |
eval=True,
|
| 88 |
use_deepspeed=False,
|
| 89 |
)
|
| 90 |
-
#model.cuda()
|
| 91 |
|
| 92 |
# This is for debugging purposes only
|
| 93 |
DEVICE_ASSERT_DETECTED = 0
|
|
@@ -683,19 +680,17 @@ with gr.Blocks(analytics_enabled=False) as demo:
|
|
| 683 |
"hu",
|
| 684 |
"hi"
|
| 685 |
],
|
| 686 |
-
|
| 687 |
value="et",
|
| 688 |
)
|
| 689 |
ref_gr = gr.Audio(
|
| 690 |
label="Reference Audio",
|
| 691 |
-
info="Click on the ✎ button to upload your own target speaker audio",
|
| 692 |
type="filepath",
|
| 693 |
value="examples/LJ001-0030.wav",
|
| 694 |
)
|
| 695 |
mic_gr = gr.Audio(
|
| 696 |
source="microphone",
|
| 697 |
type="filepath",
|
| 698 |
-
info="Use your microphone to record audio",
|
| 699 |
label="Use Microphone for Reference",
|
| 700 |
)
|
| 701 |
use_mic_gr = gr.Checkbox(
|
|
|
|
| 50 |
# This will trigger downloading model
|
| 51 |
print("Downloading Coqui XTTS V2")
|
| 52 |
|
|
|
|
|
|
|
| 53 |
model_folder = "model/"
|
| 54 |
config_file = "config.json"
|
| 55 |
if not os.path.exists(model_folder + config_file):
|
|
|
|
| 85 |
eval=True,
|
| 86 |
use_deepspeed=False,
|
| 87 |
)
|
|
|
|
| 88 |
|
| 89 |
# This is for debugging purposes only
|
| 90 |
DEVICE_ASSERT_DETECTED = 0
|
|
|
|
| 680 |
"hu",
|
| 681 |
"hi"
|
| 682 |
],
|
| 683 |
+
multiselect=False,
|
| 684 |
value="et",
|
| 685 |
)
|
| 686 |
ref_gr = gr.Audio(
|
| 687 |
label="Reference Audio",
|
|
|
|
| 688 |
type="filepath",
|
| 689 |
value="examples/LJ001-0030.wav",
|
| 690 |
)
|
| 691 |
mic_gr = gr.Audio(
|
| 692 |
source="microphone",
|
| 693 |
type="filepath",
|
|
|
|
| 694 |
label="Use Microphone for Reference",
|
| 695 |
)
|
| 696 |
use_mic_gr = gr.Checkbox(
|
app_local.py
CHANGED
|
@@ -21,7 +21,6 @@ def load_model():
|
|
| 21 |
return XTTS_MODEL
|
| 22 |
|
| 23 |
model = load_model()
|
| 24 |
-
#model.cuda()
|
| 25 |
|
| 26 |
def predict(sentence, language, reference_clip):
|
| 27 |
if not reference_clip or not reference_clip.split('.')[-1] in ['mp3', 'wav']:
|
|
|
|
| 21 |
return XTTS_MODEL
|
| 22 |
|
| 23 |
model = load_model()
|
|
|
|
| 24 |
|
| 25 |
def predict(sentence, language, reference_clip):
|
| 26 |
if not reference_clip or not reference_clip.split('.')[-1] in ['mp3', 'wav']:
|
requirements.txt
CHANGED
|
@@ -27,7 +27,6 @@ pandas>=1.4,<2.0
|
|
| 27 |
# deps for training
|
| 28 |
matplotlib>=3.7.0
|
| 29 |
# coqui stack
|
| 30 |
-
trainer>=0.0.36
|
| 31 |
coqui-tts-trainer>=0.2.0,<0.3.0
|
| 32 |
# config management
|
| 33 |
coqpit-config>=0.1.1,<0.2.0
|
|
|
|
| 27 |
# deps for training
|
| 28 |
matplotlib>=3.7.0
|
| 29 |
# coqui stack
|
|
|
|
| 30 |
coqui-tts-trainer>=0.2.0,<0.3.0
|
| 31 |
# config management
|
| 32 |
coqpit-config>=0.1.1,<0.2.0
|