[ { "TestClass": "AlbertTokenizationTest", "sp_base": false, "sequence": "I was born in 92000, and this is falsé.", "tokens": [ "▁i", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁false", "." ], "encoded": [ 31, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 4997, 9 ], "encoded_special": [ 2, 31, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 4997, 9, 3 ], "params": {}, "params_encode": {} }, { "TestClass": "BertTokenizationTest", "sp_base": false, "sequence": "UNwantéd,running", "tokens": [ "un", "##want", "##ed", ",", "runn", "##ing" ], "encoded": [ 9, 6, 7, 12, 10, 11 ], "encoded_special": [ 1, 9, 6, 7, 12, 10, 11, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "BertTokenizationTest", "sp_base": false, "sequence": "UNwantéd,running", "tokens": [ "un", "##want", "##ed", ",", "runn", "##ing" ], "encoded": [ 9, 6, 7, 12, 10, 11 ], "encoded_special": [ 1, 9, 6, 7, 12, 10, 11, 2 ], "params": { "do_lower_case": true }, "params_encode": {} }, { "TestClass": "BigBirdPegasusTokenizationTest", "sp_base": true, "sequence": "This is a test", "tokens": [ "▁This", "▁is", "▁a", "▁", "t", "est" ], "encoded": [ 288, 46, 9, 3, 12, 390 ], "encoded_special": [ 288, 46, 9, 3, 12, 390, 1 ], "params": {}, "params_encode": {} }, { "TestClass": "BigBirdTokenizationTest", "sp_base": false, "sequence": "I was born in 92000, and this is falsé.", "tokens": [ "▁I", "▁was", "▁b", "or", "n", "▁in", "▁", "9", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "é", "." ], "encoded": [ 8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4 ], "encoded_special": [ 1002, 8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4, 1000 ], "params": {}, "params_encode": {} }, { "TestClass": "CLIPTokenizationTest", "sp_base": false, "sequence": "lower newer", "tokens": [ "lo", "w", "er", "n", "e", "w", "er" ], "encoded": [ 10, 2, 16, 9, 3, 2, 16 ], "encoded_special": [ 21, 10, 2, 16, 9, 3, 2, 16, 22 ], "params": {}, "params_encode": {} }, { "TestClass": "CamembertTokenizationTest", "sp_base": true, "sequence": "I was born in 92000, and this is falsé.", "tokens": [ "▁I", "▁was", "▁b", "or", "n", "▁in", "▁", "9", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "é", "." ], "encoded": [ 12, 25, 88, 59, 28, 23, 11, 3, 606, 351, 351, 351, 7, 16, 70, 50, 76, 84, 10, 3, 8 ], "encoded_special": [ 5, 12, 25, 88, 59, 28, 23, 11, 3, 606, 351, 351, 351, 7, 16, 70, 50, 76, 84, 10, 3, 8, 6 ], "params": {}, "params_encode": {} }, { "TestClass": "CodeGenTokenizationTest", "sp_base": false, "sequence": "lower newer", "tokens": [ "Ġlow", "er", "Ġ", "n", "e", "w", "er" ], "encoded": [ 14, 15, 10, 9, 3, 2, 15 ], "encoded_special": [ 14, 15, 10, 9, 3, 2, 15 ], "params": { "add_prefix_space": true }, "params_encode": {} }, { "TestClass": "DPRContextEncoderTokenizationTest", "sp_base": false, "sequence": "UNwantéd,running", "tokens": [ "un", "##want", "##ed", ",", "runn", "##ing" ], "encoded": [ 9, 6, 7, 12, 10, 11 ], "encoded_special": [ 1, 9, 6, 7, 12, 10, 11, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "DPRQuestionEncoderTokenizationTest", "sp_base": false, "sequence": "UNwantéd,running", "tokens": [ "un", "##want", "##ed", ",", "runn", "##ing" ], "encoded": [ 9, 6, 7, 12, 10, 11 ], "encoded_special": [ 1, 9, 6, 7, 12, 10, 11, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "DPRReaderTokenizationTest", "sp_base": false, "sequence": "UNwantéd,running", "tokens": [ "un", "##want", "##ed", ",", "runn", "##ing" ], "encoded": [ 9, 6, 7, 12, 10, 11 ], "encoded_special": [ 1, 9, 6, 7, 12, 10, 11, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "DebertaTokenizationTest", "sp_base": false, "sequence": "lower newer", "tokens": [ "l", "o", "w", "er", "Ġ", "n", "e", "w", "er" ], "encoded": [ 0, 1, 2, 15, 10, 9, 3, 2, 15 ], "encoded_special": [ 20, 0, 1, 2, 15, 10, 9, 3, 2, 15, 21 ], "params": {}, "params_encode": {} }, { "TestClass": "DebertaV2TokenizationTest", "sp_base": false, "sequence": "I was born in 92000, and this is falsé!", "tokens": [ "▁", "I", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁fal", "s", "é", "!" ], "encoded": [ 13, 1, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 187 ], "encoded_special": [ 2, 13, 1, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 187, 3 ], "params": {}, "params_encode": {} }, { "TestClass": "DistilBertTokenizationTest", "sp_base": false, "sequence": "UNwantéd,running", "tokens": [ "un", "##want", "##ed", ",", "runn", "##ing" ], "encoded": [ 9, 6, 7, 12, 10, 11 ], "encoded_special": [ 1, 9, 6, 7, 12, 10, 11, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "ElectraTokenizationTest", "sp_base": false, "sequence": "UNwantéd,running", "tokens": [ "un", "##want", "##ed", ",", "runn", "##ing" ], "encoded": [ 9, 6, 7, 12, 10, 11 ], "encoded_special": [ 1, 9, 6, 7, 12, 10, 11, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "ElectraTokenizationTest", "sp_base": false, "sequence": "UNwantéd,running", "tokens": [ "un", "##want", "##ed", ",", "runn", "##ing" ], "encoded": [ 9, 6, 7, 12, 10, 11 ], "encoded_special": [ 1, 9, 6, 7, 12, 10, 11, 2 ], "params": { "do_lower_case": true }, "params_encode": {} }, { "TestClass": "FNetTokenizationTest", "sp_base": false, "sequence": "I was born in 92000, and this is falsé.", "tokens": [ "▁", "I", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁fal", "s", "é", "." ], "encoded": [ 13, 1, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9 ], "encoded_special": [ 2, 13, 1, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9, 3 ], "params": {}, "params_encode": {} }, { "TestClass": "FunnelTokenizationTest", "sp_base": false, "sequence": "UNwantéd,running", "tokens": [ "un", "##want", "##ed", ",", "runn", "##ing" ], "encoded": [ 7, 4, 5, 10, 8, 9 ], "encoded_special": [ 1, 7, 4, 5, 10, 8, 9, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "GPT2TokenizationTest", "sp_base": false, "sequence": "lower newer", "tokens": [ "Ġlow", "er", "Ġ", "n", "e", "w", "er" ], "encoded": [ 14, 15, 10, 9, 3, 2, 15 ], "encoded_special": [ 14, 15, 10, 9, 3, 2, 15 ], "params": { "add_prefix_space": true }, "params_encode": {} }, { "TestClass": "HerbertTokenizationTest", "sp_base": false, "sequence": "lower,newer", "tokens": [ "low", "er", ",", "n", "e", "w", "er" ], "encoded": [ 16, 17, 22, 11, 5, 4, 17 ], "encoded_special": [ 0, 16, 17, 22, 11, 5, 4, 17, 1 ], "params": {}, "params_encode": {} }, { "TestClass": "LayoutLMTokenizationTest", "sp_base": false, "sequence": "UNwantéd,running", "tokens": [ "un", "##want", "##ed", ",", "runn", "##ing" ], "encoded": [ 7, 4, 5, 10, 8, 9 ], "encoded_special": [ 1, 7, 4, 5, 10, 8, 9, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "LayoutLMv2TokenizationTest", "sp_base": false, "sequence": null, "tokens": null, "encoded": [ 10, 11, 12, 13 ], "encoded_special": [ 1, 10, 11, 12, 13, 2 ], "params": {}, "params_encode": { "text": [ "a", "weirdly", "test" ], "boxes": [ [ 423, 237, 440, 251 ], [ 427, 272, 441, 287 ], [ 419, 115, 437, 129 ] ] } }, { "TestClass": "LayoutLMv3TokenizationTest", "sp_base": false, "sequence": null, "tokens": null, "encoded": [ 14, 15, 10, 9, 3, 2, 15 ], "encoded_special": [ 20, 14, 15, 10, 9, 3, 2, 15, 21 ], "params": {}, "params_encode": { "text": [ "lower", "newer" ], "boxes": [ [ 423, 237, 440, 251 ], [ 427, 272, 441, 287 ] ] } }, { "TestClass": "LayoutXLMTokenizationTest", "sp_base": true, "sequence": null, "tokens": null, "encoded": [ 11, 113, 159, 17, 39, 171, 383 ], "encoded_special": [ 0, 11, 113, 159, 17, 39, 171, 383, 2 ], "params": {}, "params_encode": { "text": [ "a", "weirdly", "test" ], "boxes": [ [ 423, 237, 440, 251 ], [ 427, 272, 441, 287 ], [ 419, 115, 437, 129 ] ] } }, { "TestClass": "LongformerTokenizationTest", "sp_base": false, "sequence": "lower newer", "tokens": [ "l", "o", "w", "er", "Ġ", "n", "e", "w", "er" ], "encoded": [ 0, 1, 2, 15, 10, 9, 3, 2, 15 ], "encoded_special": [ 20, 0, 1, 2, 15, 10, 9, 3, 2, 15, 21 ], "params": {}, "params_encode": {} }, { "TestClass": "LxmertTokenizationTest", "sp_base": false, "sequence": "I was born in 92000, and this is falsé.", "tokens": [ "[UNK]", "[UNK]", "[UNK]", "[UNK]", "[UNK]", ",", "[UNK]", "[UNK]", "[UNK]", "[UNK]", "[UNK]" ], "encoded": [ 0, 0, 0, 0, 0, 10, 0, 0, 0, 0, 0 ], "encoded_special": [ 1, 0, 0, 0, 0, 0, 10, 0, 0, 0, 0, 0, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "MBart50TokenizationTest", "sp_base": true, "sequence": "the I to a and of in was it me that be he for with my not is s you", "tokens": [ "▁the", "▁I", "▁to", "▁a", "▁and", "▁of", "▁in", "▁was", "▁it", "▁me", "▁that", "▁be", "▁he", "▁for", "▁with", "▁my", "▁not", "▁is", "▁s", "▁you" ], "encoded": [ 6, 9, 10, 11, 13, 14, 20, 22, 26, 32, 35, 37, 38, 41, 42, 44, 45, 47, 48, 49 ], "encoded_special": [ 1004, 6, 9, 10, 11, 13, 14, 20, 22, 26, 32, 35, 37, 38, 41, 42, 44, 45, 47, 48, 49, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "MBartTokenizationTest", "sp_base": true, "sequence": "the I to a and of in was it me that be he for with my not is s you", "tokens": [ "▁the", "▁I", "▁to", "▁a", "▁and", "▁of", "▁in", "▁was", "▁it", "▁me", "▁that", "▁be", "▁he", "▁for", "▁with", "▁my", "▁not", "▁is", "▁s", "▁you" ], "encoded": [ 6, 9, 10, 11, 13, 14, 20, 22, 26, 32, 35, 37, 38, 41, 42, 44, 45, 47, 48, 49 ], "encoded_special": [ 6, 9, 10, 11, 13, 14, 20, 22, 26, 32, 35, 37, 38, 41, 42, 44, 45, 47, 48, 49, 2, 1004 ], "params": {}, "params_encode": {} }, { "TestClass": "MPNetTokenizerTest", "sp_base": false, "sequence": "UNwantéd,running", "tokens": [ "un", "##want", "##ed", ",", "runn", "##ing" ], "encoded": [ 9, 6, 7, 12, 10, 11 ], "encoded_special": [ 15, 9, 6, 7, 12, 10, 11, 16 ], "params": {}, "params_encode": {} }, { "TestClass": "MarkupLMTokenizationTest", "sp_base": false, "sequence": null, "tokens": null, "encoded": [ 21, 3, 0, 0, 1, 2, 1, 4, 0, 8 ], "encoded_special": [ 22, 21, 3, 0, 0, 1, 2, 1, 4, 0, 8, 23 ], "params": {}, "params_encode": { "text": [ "hello", "world" ], "xpaths": [ ",/html/body/div/li[1]/div/span", ",/html/body/div/li[1]/div/span" ] } }, { "TestClass": "MobileBERTTokenizationTest", "sp_base": false, "sequence": "UNwantéd,running", "tokens": [ "un", "##want", "##ed", ",", "runn", "##ing" ], "encoded": [ 9, 6, 7, 12, 10, 11 ], "encoded_special": [ 1, 9, 6, 7, 12, 10, 11, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "NllbTokenizationTest", "sp_base": true, "sequence": "the I to a and of in was it me that be he for with my not is s you", "tokens": [ "▁the", "▁I", "▁to", "▁a", "▁and", "▁of", "▁in", "▁was", "▁it", "▁me", "▁that", "▁be", "▁he", "▁for", "▁with", "▁my", "▁not", "▁is", "▁s", "▁you" ], "encoded": [ 6, 9, 10, 11, 13, 14, 20, 22, 26, 32, 35, 37, 38, 41, 42, 44, 45, 47, 48, 49 ], "encoded_special": [ 1048, 6, 9, 10, 11, 13, 14, 20, 22, 26, 32, 35, 37, 38, 41, 42, 44, 45, 47, 48, 49, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "OpenAIGPTTokenizationTest", "sp_base": false, "sequence": "lower newer", "tokens": [ "low", "er", "n", "e", "w", "er" ], "encoded": [ 14, 15, 9, 3, 2, 15 ], "encoded_special": [ 14, 15, 9, 3, 2, 15 ], "params": {}, "params_encode": {} }, { "TestClass": "PegasusTokenizationTest", "sp_base": true, "sequence": "This is a test", "tokens": [ "▁This", "▁is", "▁a", "▁", "t", "est" ], "encoded": [ 391, 149, 112, 106, 115, 493 ], "encoded_special": [ 391, 149, 112, 106, 115, 493, 1 ], "params": {}, "params_encode": {} }, { "TestClass": "Qwen2TokenizationTest", "sp_base": false, "sequence": "lower lower newer 010;}\r\n<|endoftext|>ϓ", "tokens": [ "l", "o", "w", "er", "Ġlow", "er", "Ġ", "n", "e", "w", "er", "Ġ", "0", "1", "0", ";}", "č", "Ċ", "<|endoftext|>", "Ïĵ" ], "encoded": [ 75, 78, 86, 260, 259, 260, 220, 77, 68, 86, 260, 220, 15, 16, 15, 265, 201, 198, 270, 267 ], "encoded_special": [ 75, 78, 86, 260, 259, 260, 220, 77, 68, 86, 260, 220, 15, 16, 15, 265, 201, 198, 270, 267 ], "params": {}, "params_encode": {} }, { "TestClass": "ReformerTokenizationTest", "sp_base": false, "sequence": "I was born in 92000, and this is falsé.", "tokens": [ "▁I", "▁was", "▁b", "or", "n", "▁in", "▁", "9", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "é", "." ], "encoded": [ 8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4 ], "encoded_special": [ 8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4 ], "params": {}, "params_encode": {} }, { "TestClass": "RemBertTokenizationTest", "sp_base": true, "sequence": "this is a test", "tokens": [ "▁this", "▁is", "▁a", "▁t", "est" ], "encoded": [ 66, 46, 10, 170, 382 ], "encoded_special": [ 1000, 66, 46, 10, 170, 382, 1001 ], "params": {}, "params_encode": {} }, { "TestClass": "RobertaTokenizationTest", "sp_base": false, "sequence": "lower newer", "tokens": [ "l", "o", "w", "er", "Ġ", "n", "e", "w", "er" ], "encoded": [ 0, 1, 2, 15, 10, 9, 3, 2, 15 ], "encoded_special": [ 20, 0, 1, 2, 15, 10, 9, 3, 2, 15, 21 ], "params": {}, "params_encode": {} }, { "TestClass": "SeamlessM4TTokenizationTest", "sp_base": true, "sequence": "the I to a and of in was it me that be he for with my not is s you", "tokens": [ "▁the", "▁I", "▁to", "▁a", "▁and", "▁of", "▁in", "▁was", "▁it", "▁me", "▁that", "▁be", "▁he", "▁for", "▁with", "▁my", "▁not", "▁is", "▁s", "▁you" ], "encoded": [ 6, 9, 10, 11, 13, 14, 20, 22, 26, 32, 35, 37, 38, 41, 42, 44, 45, 47, 48, 49 ], "encoded_special": [ 3, 1, 6, 9, 10, 11, 13, 14, 20, 22, 26, 32, 35, 37, 38, 41, 42, 44, 45, 47, 48, 49, 3 ], "params": {}, "params_encode": {} }, { "TestClass": "SplinterTokenizationTest", "sp_base": false, "sequence": "I need to test this rigor", "tokens": [ "[UNK]", "[UNK]", "[UNK]", "test", "this", "rigor" ], "encoded": [ 3, 10, 10, 10, 16, 13, 21, 1 ], "encoded_special": [ 3, 10, 10, 10, 16, 13, 21, 1 ], "params": {}, "params_encode": {} }, { "TestClass": "SqueezeBertTokenizationTest", "sp_base": false, "sequence": "UNwantéd,running", "tokens": [ "un", "##want", "##ed", ",", "runn", "##ing" ], "encoded": [ 9, 6, 7, 12, 10, 11 ], "encoded_special": [ 1, 9, 6, 7, 12, 10, 11, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "T5TokenizationTest", "sp_base": false, "sequence": "I was born in 92000, and this is falsé.", "tokens": [ "▁I", "▁was", "▁b", "or", "n", "▁in", "▁", "9", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "é", "." ], "encoded": [ 8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4 ], "encoded_special": [ 8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "TestTokenizationBart", "sp_base": false, "sequence": "lower newer", "tokens": [ "l", "o", "w", "er", "Ġ", "n", "e", "w", "er" ], "encoded": [ 0, 1, 2, 15, 10, 9, 3, 2, 15 ], "encoded_special": [ 20, 0, 1, 2, 15, 10, 9, 3, 2, 15, 21 ], "params": {}, "params_encode": {} }, { "TestClass": "TestTokenizationLED", "sp_base": false, "sequence": "lower newer", "tokens": [ "l", "o", "w", "er", "Ġ", "n", "e", "w", "er" ], "encoded": [ 0, 1, 2, 15, 10, 9, 3, 2, 15 ], "encoded_special": [ 20, 0, 1, 2, 15, 10, 9, 3, 2, 15, 21 ], "params": {}, "params_encode": {} }, { "TestClass": "TestTokenizationMvp", "sp_base": false, "sequence": "lower newer", "tokens": [ "l", "o", "w", "er", "Ġ", "n", "e", "w", "er" ], "encoded": [ 0, 1, 2, 15, 10, 9, 3, 2, 15 ], "encoded_special": [ 20, 0, 1, 2, 15, 10, 9, 3, 2, 15, 21 ], "params": {}, "params_encode": {} }, { "TestClass": "UdopTokenizationTest", "sp_base": true, "sequence": null, "tokens": null, "encoded": [ 10, 112, 158, 16, 38, 170, 382, 37, 86, 20 ], "encoded_special": [ 10, 112, 158, 16, 38, 170, 382, 37, 86, 20, 2 ], "params": {}, "params_encode": { "text": [ "a", "weirdly", "test", "hello" ], "boxes": [ [ 423, 237, 440, 251 ], [ 427, 272, 441, 287 ], [ 419, 115, 437, 129 ], [ 961, 885, 992, 912 ] ] } }, { "TestClass": "WhisperTokenizerTest", "sp_base": false, "sequence": "A BCDEFGHIJKLMNOPQRST", "tokens": [ "A", "ĠBC", "DE", "F", "GH", "I", "J", "K", "L", "M", "N", "OP", "Q", "R", "ST" ], "encoded": [ 32, 14359, 22296, 37, 4269, 40, 41, 42, 43, 44, 45, 12059, 48, 49, 6840 ], "encoded_special": [ 50258, 50363, 32, 14359, 22296, 37, 4269, 40, 41, 42, 43, 44, 45, 12059, 48, 49, 6840, 50257 ], "params": {}, "params_encode": {} }, { "TestClass": "XGLMTokenizationTest", "sp_base": false, "sequence": "I was born in 92000, and this is falsé.", "tokens": [ "▁I", "▁was", "▁b", "or", "n", "▁in", "▁", "9", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "é", "." ], "encoded": [ 9, 22, 85, 56, 25, 20, 8, 3, 603, 348, 348, 348, 4, 13, 67, 47, 73, 81, 7, 3, 5 ], "encoded_special": [ 2, 9, 22, 85, 56, 25, 20, 8, 3, 603, 348, 348, 348, 4, 13, 67, 47, 73, 81, 7, 3, 5 ], "params": {}, "params_encode": {} }, { "TestClass": "XLMRobertaTokenizationTest", "sp_base": false, "sequence": "I was born in 92000, and this is falsé.", "tokens": [ "▁I", "▁was", "▁b", "or", "n", "▁in", "▁", "9", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "é", "." ], "encoded": [ 9, 22, 85, 56, 25, 20, 8, 3, 603, 348, 348, 348, 4, 13, 67, 47, 73, 81, 7, 3, 5 ], "encoded_special": [ 0, 9, 22, 85, 56, 25, 20, 8, 3, 603, 348, 348, 348, 4, 13, 67, 47, 73, 81, 7, 3, 5, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "XLNetTokenizationTest", "sp_base": true, "sequence": "the I to a and of in was it me that be he for with my not is s you", "tokens": [ "▁the", "▁I", "▁to", "▁a", "▁and", "▁of", "▁in", "▁was", "▁it", "▁me", "▁that", "▁be", "▁he", "▁for", "▁with", "▁my", "▁not", "▁is", "▁s", "▁you" ], "encoded": [ 5, 8, 9, 10, 12, 13, 19, 21, 25, 31, 34, 36, 37, 40, 41, 43, 44, 46, 47, 48 ], "encoded_special": [ 5, 8, 9, 10, 12, 13, 19, 21, 25, 31, 34, 36, 37, 40, 41, 43, 44, 46, 47, 48, 1000, 1002 ], "params": {}, "params_encode": {} }, { "TestClass": "CamembertTokenizationTest", "sp_base": true, "sequence": "This is a test", "tokens": [ "▁This", "▁is", "▁a", "▁t", "est" ], "encoded": [ 289, 50, 14, 174, 386 ], "encoded_special": [ 5, 289, 50, 14, 174, 386, 6 ], "params": {}, "params_encode": {} }, { "TestClass": "CamembertTokenizationTest", "sp_base": true, "sequence": "the I to a and of in was it me that be he for with my not is s you", "tokens": [ "▁the", "▁I", "▁to", "▁a", "▁and", "▁of", "▁in", "▁was", "▁it", "▁me", "▁that", "▁be", "▁he", "▁for", "▁with", "▁my", "▁not", "▁is", "▁s", "▁you" ], "encoded": [ 9, 12, 13, 14, 16, 17, 23, 25, 29, 35, 38, 40, 41, 44, 45, 47, 48, 50, 51, 52 ], "encoded_special": [ 5, 9, 12, 13, 14, 16, 17, 23, 25, 29, 35, 38, 40, 41, 44, 45, 47, 48, 50, 51, 52, 6 ], "params": {}, "params_encode": {} }, { "TestClass": "MBartTokenizationTest", "sp_base": true, "sequence": "This is a test", "tokens": [ "▁This", "▁is", "▁a", "▁t", "est" ], "encoded": [ 286, 47, 11, 171, 383 ], "encoded_special": [ 286, 47, 11, 171, 383, 2, 1004 ], "params": {}, "params_encode": {} }, { "TestClass": "MBartTokenizationTest", "sp_base": true, "sequence": "I was born in 92000, and this is falsé.", "tokens": [ "▁I", "▁was", "▁b", "or", "n", "▁in", "▁", "9", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "é", "." ], "encoded": [ 9, 22, 85, 56, 25, 20, 8, 3, 603, 348, 348, 348, 4, 13, 67, 47, 73, 81, 7, 3, 5 ], "encoded_special": [ 9, 22, 85, 56, 25, 20, 8, 3, 603, 348, 348, 348, 4, 13, 67, 47, 73, 81, 7, 3, 5, 2, 1004 ], "params": {}, "params_encode": {} }, { "TestClass": "MBart50TokenizationTest", "sp_base": true, "sequence": "This is a test", "tokens": [ "▁This", "▁is", "▁a", "▁t", "est" ], "encoded": [ 286, 47, 11, 171, 383 ], "encoded_special": [ 1004, 286, 47, 11, 171, 383, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "MBart50TokenizationTest", "sp_base": true, "sequence": "I was born in 92000, and this is falsé.", "tokens": [ "▁I", "▁was", "▁b", "or", "n", "▁in", "▁", "9", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "é", "." ], "encoded": [ 9, 22, 85, 56, 25, 20, 8, 3, 603, 348, 348, 348, 4, 13, 67, 47, 73, 81, 7, 3, 5 ], "encoded_special": [ 1004, 9, 22, 85, 56, 25, 20, 8, 3, 603, 348, 348, 348, 4, 13, 67, 47, 73, 81, 7, 3, 5, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "NllbTokenizationTest", "sp_base": true, "sequence": "This is a test", "tokens": [ "▁This", "▁is", "▁a", "▁t", "est" ], "encoded": [ 286, 47, 11, 171, 383 ], "encoded_special": [ 1048, 286, 47, 11, 171, 383, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "NllbTokenizationTest", "sp_base": true, "sequence": "I was born in 92000, and this is falsé.", "tokens": [ "▁I", "▁was", "▁b", "or", "n", "▁in", "▁", "9", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "é", "." ], "encoded": [ 9, 22, 85, 56, 25, 20, 8, 3, 603, 348, 348, 348, 4, 13, 67, 47, 73, 81, 7, 3, 5 ], "encoded_special": [ 1048, 9, 22, 85, 56, 25, 20, 8, 3, 603, 348, 348, 348, 4, 13, 67, 47, 73, 81, 7, 3, 5, 2 ], "params": {}, "params_encode": {} }, { "TestClass": "RemBertTokenizationTest", "sp_base": true, "sequence": "This is a test", "tokens": [ "▁This", "▁is", "▁a", "▁t", "est" ], "encoded": [ 285, 46, 10, 170, 382 ], "encoded_special": [ 1000, 285, 46, 10, 170, 382, 1001 ], "params": {}, "params_encode": {} }, { "TestClass": "RemBertTokenizationTest", "sp_base": true, "sequence": "I was born in 92000, and this is falsé.", "tokens": [ "▁I", "▁was", "▁b", "or", "n", "▁in", "▁", "9", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "é", "." ], "encoded": [ 8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4 ], "encoded_special": [ 1000, 8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4, 1001 ], "params": {}, "params_encode": {} }, { "TestClass": "RemBertTokenizationTest", "sp_base": true, "sequence": "the I to a and of in was it me that be he for with my not is s you", "tokens": [ "▁the", "▁I", "▁to", "▁a", "▁and", "▁of", "▁in", "▁was", "▁it", "▁me", "▁that", "▁be", "▁he", "▁for", "▁with", "▁my", "▁not", "▁is", "▁s", "▁you" ], "encoded": [ 5, 8, 9, 10, 12, 13, 19, 21, 25, 31, 34, 36, 37, 40, 41, 43, 44, 46, 47, 48 ], "encoded_special": [ 1000, 5, 8, 9, 10, 12, 13, 19, 21, 25, 31, 34, 36, 37, 40, 41, 43, 44, 46, 47, 48, 1001 ], "params": {}, "params_encode": {} }, { "TestClass": "SeamlessM4TTokenizationTest", "sp_base": true, "sequence": "This is a test", "tokens": [ "▁This", "▁is", "▁a", "▁t", "est" ], "encoded": [ 286, 47, 11, 171, 383 ], "encoded_special": [ 3, 1, 286, 47, 11, 171, 383, 3 ], "params": {}, "params_encode": {} }, { "TestClass": "SeamlessM4TTokenizationTest", "sp_base": true, "sequence": "I was born in 92000, and this is falsé.", "tokens": [ "▁I", "▁was", "▁b", "or", "n", "▁in", "▁", "9", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "é", "." ], "encoded": [ 9, 22, 85, 56, 25, 20, 8, 1, 603, 348, 348, 348, 4, 13, 67, 47, 73, 81, 7, 1, 5 ], "encoded_special": [ 3, 1, 9, 22, 85, 56, 25, 20, 8, 1, 603, 348, 348, 348, 4, 13, 67, 47, 73, 81, 7, 1, 5, 3 ], "params": {}, "params_encode": {} }, { "TestClass": "XLNetTokenizationTest", "sp_base": true, "sequence": "I was born in 92000, and this is falsé.", "tokens": [ "▁I", "▁was", "▁b", "or", "n", "▁in", "▁", "9", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "é", "." ], "encoded": [ 8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4 ], "encoded_special": [ 8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4, 1000, 1002 ], "params": {}, "params_encode": {} } ]