ArthurZ HF Staff commited on
Commit
5c85222
Β·
verified Β·
1 Parent(s): 8f27a02

<?xml version="1.0" encoding="utf-8"?><testsuites name="pytest tests"><testsuite name="pytest" errors="0" failures="4" skipped="18" tests="243" time="28.981" timestamp="2025-11-16T22:40:27.820441+00:00" hostname="d89944057df4"><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_chat_template_save_loading" file="tests/test_tokenization_common.py" line="882" time="0.913"><skipped type="pytest.skip" message="tokenizer doesn't accept chat templates at input">/root/project/tests/test_tokenization_common.py:883: tokenizer doesn't accept chat templates at input</skipped></testcase><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_continue_final_message" file="tests/test_tokenization_common.py" line="1278" time="0.007" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_chat_template" file="tests/test_tokenization_common.py" line="821" time="0.950" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_chat_template_batched" file="tests/test_tokenization_common.py" line="924" time="0.003" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_jinja_loopcontrols" file="tests/test_tokenization_common.py" line="958" time="0.966" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_mecab_tokenizer_no_normalize" file="tests/models/bert_japanese/test_tokenization_bert_japanese.py" line="192" time="0.967" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_continue_final_message_with_decoy_earlier_message" file="tests/test_tokenization_common.py" line="1336" time="0.005" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_jinja_strftime" file="tests/test_tokenization_common.py" line="978" time="0.004" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_mecab_tokenizer_unidic" file="tests/models/bert_japanese/test_tokenization_bert_japanese.py" line="153" time="0.009" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_chat_template_dict" file="tests/test_tokenization_common.py" line="1361" time="0.004" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_encode_plus_with_padding_0" file="tests/test_tokenization_common.py" line="1992" time="1.002" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_encode_plus_with_padding_1" file="tests/test_tokenization_common.py" line="1992" time="0.003" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_continue_final_message_with_trim" file="tests/test_tokenization_common.py" line="1306" time="0.003" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_mecab_tokenizer_unidic_lite" file="tests/models/bert_japanese/test_tokenization_bert_japanese.py" line="142" time="0.002" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_chat_template_dict_saving" file="tests/test_tokenization_common.py" line="1377" time="0.008" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_full_tokenizer" file="tests/models/bert_japanese/test_tokenization_bert_japanese.py" line="98" time="0.001" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_mecab_tokenizer_with_option" file="tests/models/bert_japanese/test_tokenization_bert_japanese.py" line="178" time="0.001" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_jumanpp_tokenizer_trim_whitespace" file="tests/models/bert_japanese/test_tokenization_bert_japanese.py" line="335" time="1.058" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_mask_output" file="tests/test_tokenization_common.py" line="779" time="0.003" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_jumanpp_full_tokenizer_with_jumanpp_kwargs_trim_whitespace" file="tests/models/bert_japanese/test_tokenization_bert_japanese.py" line="344" time="0.066" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_chat_template_file_priority" file="tests/test_tokenization_common.py" line="1409" time="0.005" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_get_vocab" file="tests/test_tokenization_common.py" line="2104" time="0.003" /><testcase classname="tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest" name="test_conversion_reversible" file="tests/test_tokenization_common.py" line="2117" time="0.003"><failure message="AssertionError: 2 != 1">self = <tests.models.bert_japanese.test_tokenization_bert_japanese.BertJapaneseTokenizationTest testMethod=test_conversion_reversible>

Browse files
Files changed (1) hide show
  1. script.js +11 -12
script.js CHANGED
@@ -2,14 +2,13 @@ document.addEventListener('DOMContentLoaded', function() {
2
  // Parse XML test data
3
  const testData = {
4
  models: {
5
- 'bert-japanese': 4, // Based on failures in the XML
6
- 'clip': 1,
7
- 'openai': 1
8
  },
9
  tests: {
10
- 'test_tokenization': 3,
11
- 'test_internal_consistency': 2,
12
  'test_conversion_reversible': 2,
 
 
13
  'test_chat_template': 0,
14
  'test_pretokenized_inputs': 0
15
  },
@@ -17,20 +16,20 @@ document.addEventListener('DOMContentLoaded', function() {
17
  {
18
  type: 'Conversion Error',
19
  count: 2,
20
- models: ['bert-japanese'],
21
- details: 'AssertionError in test_conversion_reversible'
22
  },
23
  {
24
  type: 'Consistency Error',
25
  count: 2,
26
- models: ['bert-japanese'],
27
- details: 'AssertionError in test_internal_consistency'
28
  },
29
  {
30
- type: 'Tokenization Issue',
31
  count: 1,
32
- models: ['clip', 'openai'],
33
- details: 'Various tokenization failures'
34
  }
35
  ]
36
  };
 
2
  // Parse XML test data
3
  const testData = {
4
  models: {
5
+ 'bert-japanese': 4,
6
+ 'bert-japanese-character': 1
 
7
  },
8
  tests: {
 
 
9
  'test_conversion_reversible': 2,
10
+ 'test_internal_consistency': 2,
11
+ 'test_tokenization': 0,
12
  'test_chat_template': 0,
13
  'test_pretokenized_inputs': 0
14
  },
 
16
  {
17
  type: 'Conversion Error',
18
  count: 2,
19
+ models: ['bert-japanese', 'bert-japanese-character'],
20
+ details: 'AssertionError: 2 != 1 in test_conversion_reversible'
21
  },
22
  {
23
  type: 'Consistency Error',
24
  count: 2,
25
+ models: ['bert-japanese', 'bert-japanese-character'],
26
+ details: 'Text output mismatch in test_internal_consistency'
27
  },
28
  {
29
+ type: 'Truncation Error',
30
  count: 1,
31
+ models: ['bert-japanese'],
32
+ details: 'Padding/truncation issue in test_truncation_side_in_kwargs'
33
  }
34
  ]
35
  };