Commit
·
57e1a98
1
Parent(s):
818e79f
Path to pretrained model
Browse files
README.md
CHANGED
@@ -58,7 +58,7 @@ Next you can use the model using the `transformers` Python package as follows:
|
|
58 |
|
59 |
## Model Details
|
60 |
|
61 |
-
Wav2Vec2 is a state-of-the-art model architecture for speech recognition, leveraging self-supervised learning from raw audio data. The pre-trained [wav2vec2-xls-r-1b](facebook/wav2vec2-xls-r-1b) has been fine-tuned for automatic speech recognition with the [CoRal-v2 dataset](https://huggingface.co/datasets/CoRal-project/coral-v2/tree/main) dataset to enhance its performance in recognizing Danish speech with consideration to different dialects. The model was trained for 30K steps using the training setup in the [CoRaL repository](https://github.com/alexandrainst/coral/tree) by running:
|
62 |
```
|
63 |
python src/scripts/finetune_asr_model.py model=wav2vec2-medium max_steps=30000 datasets.coral_conversation_internal.id=CoRal-project/coral-v2 datasets.coral_readaloud_internal.id=CoRal-project/coral-v2
|
64 |
```
|
@@ -208,7 +208,7 @@ We would like specifically to thank Dan Saattrup Nielsen, Alexandra Institute fo
|
|
208 |
## Citation
|
209 |
```bibtex
|
210 |
@misc{roest-wav2vec2-1B-v2,
|
211 |
-
author = {Marie Juhl Jørgensen, Søren Vejlgaard Holm, Martin Carsten Nielsen, Dan Saattrup Nielsen, Sif Bernstorff Lehmann
|
212 |
title = {Roest-wav2vec-1B-v2: A Danish state-of-the-art speech recognition model trained on varied demographics and dialects},
|
213 |
year = {2025},
|
214 |
url = {https://huggingface.co/CoRal-project/roest-wav2vec2-1B-v2},
|
|
|
58 |
|
59 |
## Model Details
|
60 |
|
61 |
+
Wav2Vec2 is a state-of-the-art model architecture for speech recognition, leveraging self-supervised learning from raw audio data. The pre-trained [wav2vec2-xls-r-1b](https://huggingface.co/facebook/wav2vec2-xls-r-1b) has been fine-tuned for automatic speech recognition with the [CoRal-v2 dataset](https://huggingface.co/datasets/CoRal-project/coral-v2/tree/main) dataset to enhance its performance in recognizing Danish speech with consideration to different dialects. The model was trained for 30K steps using the training setup in the [CoRaL repository](https://github.com/alexandrainst/coral/tree) by running:
|
62 |
```
|
63 |
python src/scripts/finetune_asr_model.py model=wav2vec2-medium max_steps=30000 datasets.coral_conversation_internal.id=CoRal-project/coral-v2 datasets.coral_readaloud_internal.id=CoRal-project/coral-v2
|
64 |
```
|
|
|
208 |
## Citation
|
209 |
```bibtex
|
210 |
@misc{roest-wav2vec2-1B-v2,
|
211 |
+
author = {Marie Juhl Jørgensen, Søren Vejlgaard Holm, Martin Carsten Nielsen, Dan Saattrup Nielsen, Sif Bernstorff Lehmann, Simon Leminen Madsen and Torben Blach},
|
212 |
title = {Roest-wav2vec-1B-v2: A Danish state-of-the-art speech recognition model trained on varied demographics and dialects},
|
213 |
year = {2025},
|
214 |
url = {https://huggingface.co/CoRal-project/roest-wav2vec2-1B-v2},
|