gsaltintas commited on
Commit
951f7ca
·
verified ·
1 Parent(s): 3f91131

Uploading tokenizer_robustness_completion_english_canonical subset

Browse files
README.md CHANGED
@@ -8,7 +8,7 @@ tags:
8
  - multilingual
9
  - tokenization
10
  dataset_info:
11
- config_name: tokenizer_robustness_completion_english_abbreviations
12
  features:
13
  - name: question
14
  dtype: string
@@ -132,11 +132,139 @@ dataset_info:
132
  num_examples: 37
133
  download_size: 36121
134
  dataset_size: 19544
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
135
  configs:
136
  - config_name: tokenizer_robustness_completion_english_abbreviations
137
  data_files:
138
  - split: test
139
  path: tokenizer_robustness_completion_english_abbreviations/test-*
 
 
 
 
140
  ---
141
 
142
  # Dataset Card for Tokenization Robustness
 
8
  - multilingual
9
  - tokenization
10
  dataset_info:
11
+ - config_name: tokenizer_robustness_completion_english_abbreviations
12
  features:
13
  - name: question
14
  dtype: string
 
132
  num_examples: 37
133
  download_size: 36121
134
  dataset_size: 19544
135
+ - config_name: tokenizer_robustness_completion_english_canonical
136
+ features:
137
+ - name: question
138
+ dtype: string
139
+ - name: choices
140
+ list: string
141
+ - name: answer
142
+ dtype: int64
143
+ - name: answer_label
144
+ dtype: string
145
+ - name: split
146
+ dtype: string
147
+ - name: subcategories
148
+ dtype: string
149
+ - name: category
150
+ dtype: string
151
+ - name: lang
152
+ dtype: string
153
+ - name: second_lang
154
+ dtype: string
155
+ - name: notes
156
+ dtype: string
157
+ - name: id
158
+ dtype: string
159
+ - name: set_id
160
+ dtype: string
161
+ - name: variation_id
162
+ dtype: string
163
+ - name: vanilla_cos_sim_to_canonical
164
+ struct:
165
+ - name: CohereLabs/aya-expanse-8b
166
+ dtype: float64
167
+ - name: Qwen/Qwen3-8B
168
+ dtype: float64
169
+ - name: bigscience/bloom
170
+ dtype: float64
171
+ - name: common-pile/comma-v0.1-1t
172
+ dtype: float64
173
+ - name: facebook/xglm-564M
174
+ dtype: float64
175
+ - name: google-bert/bert-base-multilingual-cased
176
+ dtype: float64
177
+ - name: google/byt5-small
178
+ dtype: float64
179
+ - name: google/gemma-2-2b
180
+ dtype: float64
181
+ - name: gpt2
182
+ dtype: float64
183
+ - name: meta-llama/Llama-3.2-1B
184
+ dtype: float64
185
+ - name: microsoft/Phi-3-mini-4k-instruct
186
+ dtype: float64
187
+ - name: mistralai/tekken
188
+ dtype: float64
189
+ - name: tiktoken/gpt-4o
190
+ dtype: float64
191
+ - name: tokenmonster/englishcode-32000-consistent-v1
192
+ dtype: float64
193
+ - name: trimmed_cos_sim_to_canonical
194
+ struct:
195
+ - name: CohereLabs/aya-expanse-8b
196
+ dtype: float64
197
+ - name: Qwen/Qwen3-8B
198
+ dtype: float64
199
+ - name: bigscience/bloom
200
+ dtype: float64
201
+ - name: common-pile/comma-v0.1-1t
202
+ dtype: float64
203
+ - name: facebook/xglm-564M
204
+ dtype: float64
205
+ - name: google-bert/bert-base-multilingual-cased
206
+ dtype: float64
207
+ - name: google/byt5-small
208
+ dtype: float64
209
+ - name: google/gemma-2-2b
210
+ dtype: float64
211
+ - name: gpt2
212
+ dtype: float64
213
+ - name: meta-llama/Llama-3.2-1B
214
+ dtype: float64
215
+ - name: microsoft/Phi-3-mini-4k-instruct
216
+ dtype: float64
217
+ - name: mistralai/tekken
218
+ dtype: float64
219
+ - name: tiktoken/gpt-4o
220
+ dtype: float64
221
+ - name: tokenmonster/englishcode-32000-consistent-v1
222
+ dtype: float64
223
+ - name: token_counts
224
+ struct:
225
+ - name: CohereLabs/aya-expanse-8b
226
+ dtype: int64
227
+ - name: Qwen/Qwen3-8B
228
+ dtype: int64
229
+ - name: bigscience/bloom
230
+ dtype: int64
231
+ - name: common-pile/comma-v0.1-1t
232
+ dtype: int64
233
+ - name: facebook/xglm-564M
234
+ dtype: int64
235
+ - name: google-bert/bert-base-multilingual-cased
236
+ dtype: int64
237
+ - name: google/byt5-small
238
+ dtype: int64
239
+ - name: google/gemma-2-2b
240
+ dtype: int64
241
+ - name: gpt2
242
+ dtype: int64
243
+ - name: meta-llama/Llama-3.2-1B
244
+ dtype: int64
245
+ - name: microsoft/Phi-3-mini-4k-instruct
246
+ dtype: int64
247
+ - name: mistralai/tekken
248
+ dtype: int64
249
+ - name: tiktoken/gpt-4o
250
+ dtype: int64
251
+ - name: tokenmonster/englishcode-32000-consistent-v1
252
+ dtype: int64
253
+ splits:
254
+ - name: test
255
+ num_bytes: 21874
256
+ num_examples: 40
257
+ download_size: 32721
258
+ dataset_size: 21874
259
  configs:
260
  - config_name: tokenizer_robustness_completion_english_abbreviations
261
  data_files:
262
  - split: test
263
  path: tokenizer_robustness_completion_english_abbreviations/test-*
264
+ - config_name: tokenizer_robustness_completion_english_canonical
265
+ data_files:
266
+ - split: test
267
+ path: tokenizer_robustness_completion_english_canonical/test-*
268
  ---
269
 
270
  # Dataset Card for Tokenization Robustness
tokenizer_robustness_completion_english_canonical/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58c75b0aedfdc3250649d9afa938fde6ccda4e681fbff49ee8720e268ea08149
3
+ size 32721