khushi1234455687 commited on
Commit
980edc1
·
verified ·
1 Parent(s): 44de1c6

Upload tokenizer

Browse files
Files changed (5) hide show
  1. README.md +7 -7
  2. added_tokens.json +4 -0
  3. special_tokens_map.json +6 -0
  4. tokenizer_config.json +47 -0
  5. vocab.json +114 -0
README.md CHANGED
@@ -1,19 +1,19 @@
1
  ---
2
- library_name: transformers
3
- license: apache-2.0
4
  base_model: facebook/wav2vec2-large-xlsr-53
5
- tags:
6
- - generated_from_trainer
7
  datasets:
8
  - fleurs
 
 
9
  metrics:
10
  - wer
 
 
11
  model-index:
12
  - name: wav2vec2-large-xlsr-53-Hindi-Version3
13
  results:
14
  - task:
15
- name: Automatic Speech Recognition
16
  type: automatic-speech-recognition
 
17
  dataset:
18
  name: fleurs
19
  type: fleurs
@@ -21,9 +21,9 @@ model-index:
21
  split: None
22
  args: hi_in
23
  metrics:
24
- - name: Wer
25
- type: wer
26
  value: 0.282143903153373
 
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
1
  ---
 
 
2
  base_model: facebook/wav2vec2-large-xlsr-53
 
 
3
  datasets:
4
  - fleurs
5
+ library_name: transformers
6
+ license: apache-2.0
7
  metrics:
8
  - wer
9
+ tags:
10
+ - generated_from_trainer
11
  model-index:
12
  - name: wav2vec2-large-xlsr-53-Hindi-Version3
13
  results:
14
  - task:
 
15
  type: automatic-speech-recognition
16
+ name: Automatic Speech Recognition
17
  dataset:
18
  name: fleurs
19
  type: fleurs
 
21
  split: None
22
  args: hi_in
23
  metrics:
24
+ - type: wer
 
25
  value: 0.282143903153373
26
+ name: Wer
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "</s>": 113,
3
+ "<s>": 112
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "eos_token": "</s>",
4
+ "pad_token": "[PAD]",
5
+ "unk_token": "[UNK]"
6
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "110": {
4
+ "content": "[UNK]",
5
+ "lstrip": true,
6
+ "normalized": false,
7
+ "rstrip": true,
8
+ "single_word": false,
9
+ "special": false
10
+ },
11
+ "111": {
12
+ "content": "[PAD]",
13
+ "lstrip": true,
14
+ "normalized": false,
15
+ "rstrip": true,
16
+ "single_word": false,
17
+ "special": false
18
+ },
19
+ "112": {
20
+ "content": "<s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "113": {
28
+ "content": "</s>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ }
35
+ },
36
+ "bos_token": "<s>",
37
+ "clean_up_tokenization_spaces": true,
38
+ "do_lower_case": false,
39
+ "eos_token": "</s>",
40
+ "model_max_length": 1000000000000000019884624838656,
41
+ "pad_token": "[PAD]",
42
+ "replace_word_delimiter_char": " ",
43
+ "target_lang": null,
44
+ "tokenizer_class": "Wav2Vec2CTCTokenizer",
45
+ "unk_token": "[UNK]",
46
+ "word_delimiter_token": "|"
47
+ }
vocab.json ADDED
@@ -0,0 +1,114 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "$": 86,
3
+ "'": 107,
4
+ "+": 4,
5
+ "/": 73,
6
+ "0": 41,
7
+ "1": 13,
8
+ "2": 22,
9
+ "3": 16,
10
+ "4": 100,
11
+ "5": 30,
12
+ "6": 18,
13
+ "7": 24,
14
+ "8": 54,
15
+ "9": 96,
16
+ "[PAD]": 111,
17
+ "[UNK]": 110,
18
+ "a": 90,
19
+ "b": 11,
20
+ "c": 28,
21
+ "d": 62,
22
+ "e": 26,
23
+ "f": 44,
24
+ "g": 52,
25
+ "h": 109,
26
+ "i": 99,
27
+ "j": 103,
28
+ "k": 40,
29
+ "l": 19,
30
+ "m": 47,
31
+ "n": 66,
32
+ "o": 46,
33
+ "p": 108,
34
+ "q": 78,
35
+ "r": 51,
36
+ "s": 75,
37
+ "t": 57,
38
+ "u": 34,
39
+ "v": 21,
40
+ "w": 97,
41
+ "x": 15,
42
+ "y": 77,
43
+ "z": 85,
44
+ "|": 10,
45
+ "£": 76,
46
+ "¥": 25,
47
+ "°": 87,
48
+ "á": 39,
49
+ "õ": 27,
50
+ "ú": 72,
51
+ "ँ": 104,
52
+ "ं": 58,
53
+ "ः": 56,
54
+ "अ": 38,
55
+ "आ": 29,
56
+ "इ": 49,
57
+ "ई": 42,
58
+ "उ": 80,
59
+ "ऊ": 35,
60
+ "ए": 12,
61
+ "ऐ": 82,
62
+ "ऑ": 79,
63
+ "ओ": 69,
64
+ "औ": 6,
65
+ "क": 102,
66
+ "ख": 94,
67
+ "ग": 33,
68
+ "घ": 23,
69
+ "च": 64,
70
+ "छ": 98,
71
+ "ज": 55,
72
+ "झ": 2,
73
+ "ञ": 48,
74
+ "ट": 65,
75
+ "ठ": 20,
76
+ "ड": 91,
77
+ "ढ": 70,
78
+ "ण": 3,
79
+ "त": 63,
80
+ "थ": 95,
81
+ "द": 71,
82
+ "ध": 93,
83
+ "न": 60,
84
+ "प": 88,
85
+ "फ": 67,
86
+ "ब": 59,
87
+ "भ": 8,
88
+ "म": 92,
89
+ "य": 9,
90
+ "र": 61,
91
+ "ल": 14,
92
+ "व": 68,
93
+ "श": 53,
94
+ "ष": 74,
95
+ "स": 5,
96
+ "ह": 105,
97
+ "़": 31,
98
+ "ा": 81,
99
+ "ि": 7,
100
+ "ी": 36,
101
+ "ु": 83,
102
+ "ू": 0,
103
+ "ृ": 101,
104
+ "े": 32,
105
+ "ै": 106,
106
+ "ॉ": 37,
107
+ "ो": 84,
108
+ "ौ": 50,
109
+ "्": 43,
110
+ "।": 17,
111
+ "​": 89,
112
+ "‍": 45,
113
+ "—": 1
114
+ }