Datasets:
Uploading tokenizer_robustness_completion_english_capitalization subset
Browse files
README.md
CHANGED
@@ -256,6 +256,130 @@ dataset_info:
|
|
256 |
num_examples: 40
|
257 |
download_size: 32721
|
258 |
dataset_size: 21874
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
259 |
configs:
|
260 |
- config_name: tokenizer_robustness_completion_english_abbreviations
|
261 |
data_files:
|
@@ -265,6 +389,10 @@ configs:
|
|
265 |
data_files:
|
266 |
- split: test
|
267 |
path: tokenizer_robustness_completion_english_canonical/test-*
|
|
|
|
|
|
|
|
|
268 |
---
|
269 |
|
270 |
# Dataset Card for Tokenization Robustness
|
|
|
256 |
num_examples: 40
|
257 |
download_size: 32721
|
258 |
dataset_size: 21874
|
259 |
+
- config_name: tokenizer_robustness_completion_english_capitalization
|
260 |
+
features:
|
261 |
+
- name: question
|
262 |
+
dtype: string
|
263 |
+
- name: choices
|
264 |
+
list: string
|
265 |
+
- name: answer
|
266 |
+
dtype: int64
|
267 |
+
- name: answer_label
|
268 |
+
dtype: string
|
269 |
+
- name: split
|
270 |
+
dtype: string
|
271 |
+
- name: subcategories
|
272 |
+
dtype: string
|
273 |
+
- name: category
|
274 |
+
dtype: string
|
275 |
+
- name: lang
|
276 |
+
dtype: string
|
277 |
+
- name: second_lang
|
278 |
+
dtype: string
|
279 |
+
- name: notes
|
280 |
+
dtype: string
|
281 |
+
- name: id
|
282 |
+
dtype: string
|
283 |
+
- name: set_id
|
284 |
+
dtype: string
|
285 |
+
- name: variation_id
|
286 |
+
dtype: string
|
287 |
+
- name: vanilla_cos_sim_to_canonical
|
288 |
+
struct:
|
289 |
+
- name: CohereLabs/aya-expanse-8b
|
290 |
+
dtype: float64
|
291 |
+
- name: Qwen/Qwen3-8B
|
292 |
+
dtype: float64
|
293 |
+
- name: bigscience/bloom
|
294 |
+
dtype: float64
|
295 |
+
- name: common-pile/comma-v0.1-1t
|
296 |
+
dtype: float64
|
297 |
+
- name: facebook/xglm-564M
|
298 |
+
dtype: float64
|
299 |
+
- name: google-bert/bert-base-multilingual-cased
|
300 |
+
dtype: float64
|
301 |
+
- name: google/byt5-small
|
302 |
+
dtype: float64
|
303 |
+
- name: google/gemma-2-2b
|
304 |
+
dtype: float64
|
305 |
+
- name: gpt2
|
306 |
+
dtype: float64
|
307 |
+
- name: meta-llama/Llama-3.2-1B
|
308 |
+
dtype: float64
|
309 |
+
- name: microsoft/Phi-3-mini-4k-instruct
|
310 |
+
dtype: float64
|
311 |
+
- name: mistralai/tekken
|
312 |
+
dtype: float64
|
313 |
+
- name: tiktoken/gpt-4o
|
314 |
+
dtype: float64
|
315 |
+
- name: tokenmonster/englishcode-32000-consistent-v1
|
316 |
+
dtype: float64
|
317 |
+
- name: trimmed_cos_sim_to_canonical
|
318 |
+
struct:
|
319 |
+
- name: CohereLabs/aya-expanse-8b
|
320 |
+
dtype: float64
|
321 |
+
- name: Qwen/Qwen3-8B
|
322 |
+
dtype: float64
|
323 |
+
- name: bigscience/bloom
|
324 |
+
dtype: float64
|
325 |
+
- name: common-pile/comma-v0.1-1t
|
326 |
+
dtype: float64
|
327 |
+
- name: facebook/xglm-564M
|
328 |
+
dtype: float64
|
329 |
+
- name: google-bert/bert-base-multilingual-cased
|
330 |
+
dtype: float64
|
331 |
+
- name: google/byt5-small
|
332 |
+
dtype: float64
|
333 |
+
- name: google/gemma-2-2b
|
334 |
+
dtype: float64
|
335 |
+
- name: gpt2
|
336 |
+
dtype: float64
|
337 |
+
- name: meta-llama/Llama-3.2-1B
|
338 |
+
dtype: float64
|
339 |
+
- name: microsoft/Phi-3-mini-4k-instruct
|
340 |
+
dtype: float64
|
341 |
+
- name: mistralai/tekken
|
342 |
+
dtype: float64
|
343 |
+
- name: tiktoken/gpt-4o
|
344 |
+
dtype: float64
|
345 |
+
- name: tokenmonster/englishcode-32000-consistent-v1
|
346 |
+
dtype: float64
|
347 |
+
- name: token_counts
|
348 |
+
struct:
|
349 |
+
- name: CohereLabs/aya-expanse-8b
|
350 |
+
dtype: int64
|
351 |
+
- name: Qwen/Qwen3-8B
|
352 |
+
dtype: int64
|
353 |
+
- name: bigscience/bloom
|
354 |
+
dtype: int64
|
355 |
+
- name: common-pile/comma-v0.1-1t
|
356 |
+
dtype: int64
|
357 |
+
- name: facebook/xglm-564M
|
358 |
+
dtype: int64
|
359 |
+
- name: google-bert/bert-base-multilingual-cased
|
360 |
+
dtype: int64
|
361 |
+
- name: google/byt5-small
|
362 |
+
dtype: int64
|
363 |
+
- name: google/gemma-2-2b
|
364 |
+
dtype: int64
|
365 |
+
- name: gpt2
|
366 |
+
dtype: int64
|
367 |
+
- name: meta-llama/Llama-3.2-1B
|
368 |
+
dtype: int64
|
369 |
+
- name: microsoft/Phi-3-mini-4k-instruct
|
370 |
+
dtype: int64
|
371 |
+
- name: mistralai/tekken
|
372 |
+
dtype: int64
|
373 |
+
- name: tiktoken/gpt-4o
|
374 |
+
dtype: int64
|
375 |
+
- name: tokenmonster/englishcode-32000-consistent-v1
|
376 |
+
dtype: int64
|
377 |
+
splits:
|
378 |
+
- name: test
|
379 |
+
num_bytes: 21413
|
380 |
+
num_examples: 40
|
381 |
+
download_size: 40005
|
382 |
+
dataset_size: 21413
|
383 |
configs:
|
384 |
- config_name: tokenizer_robustness_completion_english_abbreviations
|
385 |
data_files:
|
|
|
389 |
data_files:
|
390 |
- split: test
|
391 |
path: tokenizer_robustness_completion_english_canonical/test-*
|
392 |
+
- config_name: tokenizer_robustness_completion_english_capitalization
|
393 |
+
data_files:
|
394 |
+
- split: test
|
395 |
+
path: tokenizer_robustness_completion_english_capitalization/test-*
|
396 |
---
|
397 |
|
398 |
# Dataset Card for Tokenization Robustness
|
tokenizer_robustness_completion_english_capitalization/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:45c30c5a021d5f0d98165a8a9ced89e0e69a671dc7ad740f42e4bb6a7221cf08
|
3 |
+
size 40005
|