gsaltintas commited on
Commit
6e89efa
·
verified ·
1 Parent(s): ef9f04d

Uploading tokenizer_robustness_completion_chinese_keyboard_proximity_errors subset

Browse files
README.md CHANGED
@@ -144,6 +144,40 @@ dataset_info:
144
  num_examples: 40
145
  download_size: 8383
146
  dataset_size: 7907
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
147
  configs:
148
  - config_name: tokenizer_robustness_completion_chinese_canonical
149
  data_files:
@@ -161,6 +195,10 @@ configs:
161
  data_files:
162
  - split: test
163
  path: tokenizer_robustness_completion_chinese_equivalent_expressions/test-*
 
 
 
 
164
  ---
165
 
166
  # Dataset Card for Tokenization Robustness
 
144
  num_examples: 40
145
  download_size: 8383
146
  dataset_size: 7907
147
+ - config_name: tokenizer_robustness_completion_chinese_keyboard_proximity_errors
148
+ features:
149
+ - name: question
150
+ dtype: string
151
+ - name: choices
152
+ list: string
153
+ - name: answer
154
+ dtype: int64
155
+ - name: answer_label
156
+ dtype: string
157
+ - name: split
158
+ dtype: string
159
+ - name: subcategories
160
+ dtype: string
161
+ - name: category
162
+ dtype: string
163
+ - name: lang
164
+ dtype: string
165
+ - name: second_lang
166
+ dtype: string
167
+ - name: notes
168
+ dtype: string
169
+ - name: id
170
+ dtype: string
171
+ - name: set_id
172
+ dtype: string
173
+ - name: variation_id
174
+ dtype: string
175
+ splits:
176
+ - name: test
177
+ num_bytes: 7340
178
+ num_examples: 40
179
+ download_size: 8251
180
+ dataset_size: 7340
181
  configs:
182
  - config_name: tokenizer_robustness_completion_chinese_canonical
183
  data_files:
 
195
  data_files:
196
  - split: test
197
  path: tokenizer_robustness_completion_chinese_equivalent_expressions/test-*
198
+ - config_name: tokenizer_robustness_completion_chinese_keyboard_proximity_errors
199
+ data_files:
200
+ - split: test
201
+ path: tokenizer_robustness_completion_chinese_keyboard_proximity_errors/test-*
202
  ---
203
 
204
  # Dataset Card for Tokenization Robustness
tokenizer_robustness_completion_chinese_keyboard_proximity_errors/test-00000-of-00001.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9eb303d1275a29bf2bb0e888918c9bf73845c4cd6e066eba03743f00e2847ee5
3
- size 7358
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:153a570894c13712e37f043a502a54c881c82a6b956561989209b12cc0b62b34
3
+ size 8251