gsaltintas commited on
Commit
5ec1d4a
·
verified ·
1 Parent(s): 7032274

Uploading tokenizer_robustness_completion_chinese_chinese-character-selection-proximity_error subset

Browse files
README.md CHANGED
@@ -76,6 +76,40 @@ dataset_info:
76
  num_examples: 40
77
  download_size: 9498
78
  dataset_size: 7872
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
79
  configs:
80
  - config_name: tokenizer_robustness_completion_chinese_borrowing
81
  data_files:
@@ -85,6 +119,10 @@ configs:
85
  data_files:
86
  - split: test
87
  path: tokenizer_robustness_completion_chinese_cannonical/test-*
 
 
 
 
88
  ---
89
 
90
  # Dataset Card for Tokenization Robustness
 
76
  num_examples: 40
77
  download_size: 9498
78
  dataset_size: 7872
79
+ - config_name: tokenizer_robustness_completion_chinese_chinese-character-selection-proximity_error
80
+ features:
81
+ - name: question
82
+ dtype: string
83
+ - name: choices
84
+ list: string
85
+ - name: answer
86
+ dtype: int64
87
+ - name: answer_label
88
+ dtype: string
89
+ - name: split
90
+ dtype: string
91
+ - name: subcategories
92
+ dtype: string
93
+ - name: category
94
+ dtype: string
95
+ - name: lang
96
+ dtype: string
97
+ - name: second_lang
98
+ dtype: string
99
+ - name: notes
100
+ dtype: string
101
+ - name: id
102
+ dtype: string
103
+ - name: set_id
104
+ dtype: float64
105
+ - name: variation_id
106
+ dtype: float64
107
+ splits:
108
+ - name: test
109
+ num_bytes: 2086
110
+ num_examples: 11
111
+ download_size: 6348
112
+ dataset_size: 2086
113
  configs:
114
  - config_name: tokenizer_robustness_completion_chinese_borrowing
115
  data_files:
 
119
  data_files:
120
  - split: test
121
  path: tokenizer_robustness_completion_chinese_cannonical/test-*
122
+ - config_name: tokenizer_robustness_completion_chinese_chinese-character-selection-proximity_error
123
+ data_files:
124
+ - split: test
125
+ path: tokenizer_robustness_completion_chinese_chinese-character-selection-proximity_error/test-*
126
  ---
127
 
128
  # Dataset Card for Tokenization Robustness
tokenizer_robustness_completion_chinese_chinese-character-selection-proximity_error/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03d32056c252aa8cbea4997a0311256a618ba3052a1d1be26bd5d231a6006201
3
+ size 6348