gsaltintas commited on
Commit
876c37c
·
verified ·
1 Parent(s): 5ec1d4a

Uploading tokenizer_robustness_completion_chinese_classifier_addition subset

Browse files
README.md CHANGED
@@ -110,6 +110,40 @@ dataset_info:
110
  num_examples: 11
111
  download_size: 6348
112
  dataset_size: 2086
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
113
  configs:
114
  - config_name: tokenizer_robustness_completion_chinese_borrowing
115
  data_files:
@@ -123,6 +157,10 @@ configs:
123
  data_files:
124
  - split: test
125
  path: tokenizer_robustness_completion_chinese_chinese-character-selection-proximity_error/test-*
 
 
 
 
126
  ---
127
 
128
  # Dataset Card for Tokenization Robustness
 
110
  num_examples: 11
111
  download_size: 6348
112
  dataset_size: 2086
113
+ - config_name: tokenizer_robustness_completion_chinese_classifier_addition
114
+ features:
115
+ - name: question
116
+ dtype: string
117
+ - name: choices
118
+ list: string
119
+ - name: answer
120
+ dtype: int64
121
+ - name: answer_label
122
+ dtype: string
123
+ - name: split
124
+ dtype: string
125
+ - name: subcategories
126
+ dtype: string
127
+ - name: category
128
+ dtype: string
129
+ - name: lang
130
+ dtype: string
131
+ - name: second_lang
132
+ dtype: string
133
+ - name: notes
134
+ dtype: string
135
+ - name: id
136
+ dtype: string
137
+ - name: set_id
138
+ dtype: float64
139
+ - name: variation_id
140
+ dtype: float64
141
+ splits:
142
+ - name: test
143
+ num_bytes: 152
144
+ num_examples: 1
145
+ download_size: 5438
146
+ dataset_size: 152
147
  configs:
148
  - config_name: tokenizer_robustness_completion_chinese_borrowing
149
  data_files:
 
157
  data_files:
158
  - split: test
159
  path: tokenizer_robustness_completion_chinese_chinese-character-selection-proximity_error/test-*
160
+ - config_name: tokenizer_robustness_completion_chinese_classifier_addition
161
+ data_files:
162
+ - split: test
163
+ path: tokenizer_robustness_completion_chinese_classifier_addition/test-*
164
  ---
165
 
166
  # Dataset Card for Tokenization Robustness
tokenizer_robustness_completion_chinese_classifier_addition/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ea0537ebf7c4be657391328ec28d0d755bb1ef46ebdb494555ccf6045559d81
3
+ size 5438