<?xml version="1.0" encoding="utf-8"?><testsuites name="pytest tests"><testsuite name="pytest" errors="0" failures="4" skipped="18" tests="243" time="28.981" timestamp="2025-11-16T22:40:27.820441+00:00" hostname="d89944057df4"><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_chat_template_save_loading" file="tests/test_tokenization_common.py" line="882" time="0.913"><skipped type="pytest.skip" message="tokenizer doesn't accept chat templates at input">/root/project/tests/test_tokenization_common.py:883: tokenizer doesn't accept chat templates at input</skipped></testcase><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_continue_final_message" file="tests/test_tokenization_common.py" line="1278" time="0.007" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_chat_template" file="tests/test_tokenization_common.py" line="821" time="0.950" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_chat_template_batched" file="tests/test_tokenization_common.py" line="924" time="0.003" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_jinja_loopcontrols" file="tests/test_tokenization_common.py" line="958" time="0.966" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_mecab_tokenizer_no_normalize" file="tests/models/bert_japanese/test_tokenization_bert_japanese.py" line="192" time="0.967" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_continue_final_message_with_decoy_earlier_message" file="tests/test_tokenization_common.py" line="1336" time="0.005" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_jinja_strftime" file="tests/test_tokenization_common.py" line="978" time="0.004" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_mecab_tokenizer_unidic" file="tests/models/bert_japanese/test_tokenization_bert_japanese.py" line="153" time="0.009" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_chat_template_dict" file="tests/test_tokenization_common.py" line="1361" time="0.004" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_encode_plus_with_padding_0" file="tests/test_tokenization_common.py" line="1992" time="1.002" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_encode_plus_with_padding_1" file="tests/test_tokenization_common.py" line="1992" time="0.003" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_continue_final_message_with_trim" file="tests/test_tokenization_common.py" line="1306" time="0.003" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_mecab_tokenizer_unidic_lite" file="tests/models/bert_japanese/test_tokenization_bert_japanese.py" line="142" time="0.002" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_chat_template_dict_saving" file="tests/test_tokenization_common.py" line="1377" time="0.008" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_full_tokenizer" file="tests/models/bert_japanese/test_tokenization_bert_japanese.py" line="98" time="0.001" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_mecab_tokenizer_with_option" file="tests/models/bert_japanese/test_tokenization_bert_japanese.py" line="178" time="0.001" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_jumanpp_tokenizer_trim_whitespace" file="tests/models/bert_japanese/test_tokenization_bert_japanese.py" line="335" time="1.058" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_mask_output" file="tests/test_tokenization_common.py" line="779" time="0.003" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_jumanpp_full_tokenizer_with_jumanpp_kwargs_trim_whitespace" file="tests/models/bert_japanese/test_tokenization_bert_japanese.py" line="344" time="0.066" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_chat_template_file_priority" file="tests/test_tokenization_common.py" line="1409" time="0.005" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_get_vocab" file="tests/test_tokenization_common.py" line="2104" time="0.003" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_conversion_reversible" file="tests/test_tokenization_common.py" line="2117" time="0.003"><failure message="AssertionError: 2 != 1">self = <tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest testMethod=test_conversion_reversible>
Browse files
script.js
CHANGED
|
@@ -2,14 +2,13 @@ document.addEventListener('DOMContentLoaded', function() {
|
|
| 2 |
// Parse XML test data
|
| 3 |
const testData = {
|
| 4 |
models: {
|
| 5 |
-
'bert-japanese': 4,
|
| 6 |
-
'
|
| 7 |
-
'openai': 1
|
| 8 |
},
|
| 9 |
tests: {
|
| 10 |
-
'test_tokenization': 3,
|
| 11 |
-
'test_internal_consistency': 2,
|
| 12 |
'test_conversion_reversible': 2,
|
|
|
|
|
|
|
| 13 |
'test_chat_template': 0,
|
| 14 |
'test_pretokenized_inputs': 0
|
| 15 |
},
|
|
@@ -17,20 +16,20 @@ document.addEventListener('DOMContentLoaded', function() {
|
|
| 17 |
{
|
| 18 |
type: 'Conversion Error',
|
| 19 |
count: 2,
|
| 20 |
-
models: ['bert-japanese'],
|
| 21 |
-
details: 'AssertionError in test_conversion_reversible'
|
| 22 |
},
|
| 23 |
{
|
| 24 |
type: 'Consistency Error',
|
| 25 |
count: 2,
|
| 26 |
-
models: ['bert-japanese'],
|
| 27 |
-
details: '
|
| 28 |
},
|
| 29 |
{
|
| 30 |
-
type: '
|
| 31 |
count: 1,
|
| 32 |
-
models: ['
|
| 33 |
-
details: '
|
| 34 |
}
|
| 35 |
]
|
| 36 |
};
|
|
|
|
| 2 |
// Parse XML test data
|
| 3 |
const testData = {
|
| 4 |
models: {
|
| 5 |
+
'bert-japanese': 4,
|
| 6 |
+
'bert-japanese-character': 1
|
|
|
|
| 7 |
},
|
| 8 |
tests: {
|
|
|
|
|
|
|
| 9 |
'test_conversion_reversible': 2,
|
| 10 |
+
'test_internal_consistency': 2,
|
| 11 |
+
'test_tokenization': 0,
|
| 12 |
'test_chat_template': 0,
|
| 13 |
'test_pretokenized_inputs': 0
|
| 14 |
},
|
|
|
|
| 16 |
{
|
| 17 |
type: 'Conversion Error',
|
| 18 |
count: 2,
|
| 19 |
+
models: ['bert-japanese', 'bert-japanese-character'],
|
| 20 |
+
details: 'AssertionError: 2 != 1 in test_conversion_reversible'
|
| 21 |
},
|
| 22 |
{
|
| 23 |
type: 'Consistency Error',
|
| 24 |
count: 2,
|
| 25 |
+
models: ['bert-japanese', 'bert-japanese-character'],
|
| 26 |
+
details: 'Text output mismatch in test_internal_consistency'
|
| 27 |
},
|
| 28 |
{
|
| 29 |
+
type: 'Truncation Error',
|
| 30 |
count: 1,
|
| 31 |
+
models: ['bert-japanese'],
|
| 32 |
+
details: 'Padding/truncation issue in test_truncation_side_in_kwargs'
|
| 33 |
}
|
| 34 |
]
|
| 35 |
};
|