Professor commited on
Commit
44a621e
·
1 Parent(s): d7ddca4
added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"<s>": 57, "</s>": 58}
runs/Nov15_12-58-48_e663056c08d5/1700053530.7602832/events.out.tfevents.1700053530.e663056c08d5.27.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93979d7b71dbfe17a37cf870b08ae83a2913b0a19bc1618aaa4c3b3358f25a67
3
+ size 4959
runs/Nov15_12-58-48_e663056c08d5/events.out.tfevents.1700053530.e663056c08d5.27.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f249b2434bb32e522122b578cfac9585ee7012493119675f728d774d5f4d4b17
3
+ size 5485
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "replace_word_delimiter_char": " ", "clean_up_tokenization_spaces": true, "model_max_length": 1000000000000000019884624838656, "processor_class": "Wav2Vec2Processor", "target_lang": null, "special_tokens_map_file": "/root/.cache/huggingface/transformers/34501271d596d47e0287fc24d6fc7acaa21d53c80b8b35b5b9b684cea652ea34.fff72a2ff0cb054ff34ffb94682df8d39f285b36438994ea9a0bc92d5cade4d5", "name_or_path": "alamsher/wav2vec2-large-xlsr-53-common-voice-sw", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
vocab.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"=": 1, "[PAD]": 56, "[UNK]": 55, "_": 2, "`": 3, "a": 4, "b": 5, "c": 6, "d": 7, "e": 8, "f": 9, "g": 10, "h": 11, "i": 12, "j": 13, "k": 14, "l": 15, "m": 16, "n": 17, "o": 18, "p": 19, "q": 20, "r": 21, "s": 22, "t": 23, "u": 24, "v": 25, "w": 26, "x": 27, "y": 28, "z": 29, "|": 0, "°": 30, "µ": 31, "á": 32, "â": 33, "ã": 34, "å": 35, "è": 36, "é": 37, "ë": 38, "í": 39, "ï": 40, "ñ": 41, "ó": 42, "ö": 43, "ø": 44, "ú": 45, "š": 46, "ū": 47, "ː": 48, "μ": 49, "י": 50, "ụ": 51, "’": 52, "•": 53, "…": 54}