alakxender commited on
Commit
1685a68
Β·
verified Β·
1 Parent(s): 27eca39

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -11,11 +11,11 @@ base_model:
11
  - facebook/mms-tts-div
12
  ---
13
 
14
- # πŸŽ™οΈ Divehi TTS – Male Voice (VITS-based)
15
 
16
  This is a fine-tuned VITS (Variational Inference with adversarial learning for end-to-end Text-to-Speech) model for Divehi speech synthesis. The model produces Male voice audio from Thaana-scripted Divehi text. Fine-tuned from Meta’s MMS-TTS architecture using a curated dataset of synthetic Divehi speech.
17
 
18
- ## 🧠 Model Details
19
 
20
  | Field | Value |
21
  |----------------------|-------------------------------------------------|
@@ -28,7 +28,7 @@ This is a fine-tuned VITS (Variational Inference with adversarial learning for e
28
  | **Inference Engine** | Transformers (πŸ€— Hugging Face) |
29
 
30
 
31
- ## πŸš€ Usage
32
 
33
  ```python
34
  from transformers import VitsModel, VitsTokenizer
@@ -44,7 +44,7 @@ waveform = model.generate(**inputs).waveform[0]
44
  torchaudio.save("output.wav", waveform.unsqueeze(0), 16000)
45
  ```
46
 
47
- ## πŸ“Š Evaluation Summary
48
 
49
  - **Model**: `alakxender/mms-tts-div-finetuned-md-m02`
50
  - **Evaluated Samples**: 3
@@ -64,7 +64,7 @@ torchaudio.save("output.wav", waveform.unsqueeze(0), 16000)
64
  - πŸ“„ Report: `outputs/report.txt`
65
  - πŸ“ˆ MOS Scores: `outputs/mos_scores.txt`
66
 
67
- ## πŸ™ Acknowledgements
68
 
69
  - [Meta MMS-TTS](https://github.com/facebookresearch/fairseq/tree/main/examples/mms)
70
  - [Tarepan's SpeechMOS](https://github.com/Tarepan/SpeechMOS)
 
11
  - facebook/mms-tts-div
12
  ---
13
 
14
+ # Divehi TTS – Male Voice (VITS-based)
15
 
16
  This is a fine-tuned VITS (Variational Inference with adversarial learning for end-to-end Text-to-Speech) model for Divehi speech synthesis. The model produces Male voice audio from Thaana-scripted Divehi text. Fine-tuned from Meta’s MMS-TTS architecture using a curated dataset of synthetic Divehi speech.
17
 
18
+ ## Model Details
19
 
20
  | Field | Value |
21
  |----------------------|-------------------------------------------------|
 
28
  | **Inference Engine** | Transformers (πŸ€— Hugging Face) |
29
 
30
 
31
+ ## Usage
32
 
33
  ```python
34
  from transformers import VitsModel, VitsTokenizer
 
44
  torchaudio.save("output.wav", waveform.unsqueeze(0), 16000)
45
  ```
46
 
47
+ ## Evaluation Summary
48
 
49
  - **Model**: `alakxender/mms-tts-div-finetuned-md-m02`
50
  - **Evaluated Samples**: 3
 
64
  - πŸ“„ Report: `outputs/report.txt`
65
  - πŸ“ˆ MOS Scores: `outputs/mos_scores.txt`
66
 
67
+ ## Acknowledgements
68
 
69
  - [Meta MMS-TTS](https://github.com/facebookresearch/fairseq/tree/main/examples/mms)
70
  - [Tarepan's SpeechMOS](https://github.com/Tarepan/SpeechMOS)