add model
Browse files
added_tokens.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"<s>": 57, "</s>": 58}
|
runs/Nov15_12-58-48_e663056c08d5/1700053530.7602832/events.out.tfevents.1700053530.e663056c08d5.27.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:93979d7b71dbfe17a37cf870b08ae83a2913b0a19bc1618aaa4c3b3358f25a67
|
| 3 |
+
size 4959
|
runs/Nov15_12-58-48_e663056c08d5/events.out.tfevents.1700053530.e663056c08d5.27.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f249b2434bb32e522122b578cfac9585ee7012493119675f728d774d5f4d4b17
|
| 3 |
+
size 5485
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "replace_word_delimiter_char": " ", "clean_up_tokenization_spaces": true, "model_max_length": 1000000000000000019884624838656, "processor_class": "Wav2Vec2Processor", "target_lang": null, "special_tokens_map_file": "/root/.cache/huggingface/transformers/34501271d596d47e0287fc24d6fc7acaa21d53c80b8b35b5b9b684cea652ea34.fff72a2ff0cb054ff34ffb94682df8d39f285b36438994ea9a0bc92d5cade4d5", "name_or_path": "alamsher/wav2vec2-large-xlsr-53-common-voice-sw", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
|
vocab.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"=": 1, "[PAD]": 56, "[UNK]": 55, "_": 2, "`": 3, "a": 4, "b": 5, "c": 6, "d": 7, "e": 8, "f": 9, "g": 10, "h": 11, "i": 12, "j": 13, "k": 14, "l": 15, "m": 16, "n": 17, "o": 18, "p": 19, "q": 20, "r": 21, "s": 22, "t": 23, "u": 24, "v": 25, "w": 26, "x": 27, "y": 28, "z": 29, "|": 0, "°": 30, "µ": 31, "á": 32, "â": 33, "ã": 34, "å": 35, "è": 36, "é": 37, "ë": 38, "í": 39, "ï": 40, "ñ": 41, "ó": 42, "ö": 43, "ø": 44, "ú": 45, "š": 46, "ū": 47, "ː": 48, "μ": 49, "י": 50, "ụ": 51, "’": 52, "•": 53, "…": 54}
|