Datasets:
Uploading tokenizer_robustness_completion_english_ocr_errors subset
Browse files
README.md
CHANGED
@@ -2364,6 +2364,130 @@ dataset_info:
|
|
2364 |
num_examples: 40
|
2365 |
download_size: 42278
|
2366 |
dataset_size: 25871
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2367 |
configs:
|
2368 |
- config_name: tokenizer_robustness_completion_english_abbreviations
|
2369 |
data_files:
|
@@ -2441,6 +2565,10 @@ configs:
|
|
2441 |
data_files:
|
2442 |
- split: test
|
2443 |
path: tokenizer_robustness_completion_english_macron_diacritic/test-*
|
|
|
|
|
|
|
|
|
2444 |
---
|
2445 |
|
2446 |
# Dataset Card for Tokenization Robustness
|
|
|
2364 |
num_examples: 40
|
2365 |
download_size: 42278
|
2366 |
dataset_size: 25871
|
2367 |
+
- config_name: tokenizer_robustness_completion_english_ocr_errors
|
2368 |
+
features:
|
2369 |
+
- name: question
|
2370 |
+
dtype: string
|
2371 |
+
- name: choices
|
2372 |
+
list: string
|
2373 |
+
- name: answer
|
2374 |
+
dtype: int64
|
2375 |
+
- name: answer_label
|
2376 |
+
dtype: string
|
2377 |
+
- name: split
|
2378 |
+
dtype: string
|
2379 |
+
- name: subcategories
|
2380 |
+
dtype: string
|
2381 |
+
- name: category
|
2382 |
+
dtype: string
|
2383 |
+
- name: lang
|
2384 |
+
dtype: string
|
2385 |
+
- name: second_lang
|
2386 |
+
dtype: string
|
2387 |
+
- name: notes
|
2388 |
+
dtype: string
|
2389 |
+
- name: id
|
2390 |
+
dtype: string
|
2391 |
+
- name: set_id
|
2392 |
+
dtype: string
|
2393 |
+
- name: variation_id
|
2394 |
+
dtype: string
|
2395 |
+
- name: vanilla_cos_sim_to_canonical
|
2396 |
+
struct:
|
2397 |
+
- name: CohereLabs/aya-expanse-8b
|
2398 |
+
dtype: float64
|
2399 |
+
- name: Qwen/Qwen3-8B
|
2400 |
+
dtype: float64
|
2401 |
+
- name: bigscience/bloom
|
2402 |
+
dtype: float64
|
2403 |
+
- name: common-pile/comma-v0.1-1t
|
2404 |
+
dtype: float64
|
2405 |
+
- name: facebook/xglm-564M
|
2406 |
+
dtype: float64
|
2407 |
+
- name: google-bert/bert-base-multilingual-cased
|
2408 |
+
dtype: float64
|
2409 |
+
- name: google/byt5-small
|
2410 |
+
dtype: float64
|
2411 |
+
- name: google/gemma-2-2b
|
2412 |
+
dtype: float64
|
2413 |
+
- name: gpt2
|
2414 |
+
dtype: float64
|
2415 |
+
- name: meta-llama/Llama-3.2-1B
|
2416 |
+
dtype: float64
|
2417 |
+
- name: microsoft/Phi-3-mini-4k-instruct
|
2418 |
+
dtype: float64
|
2419 |
+
- name: mistralai/tekken
|
2420 |
+
dtype: float64
|
2421 |
+
- name: tiktoken/gpt-4o
|
2422 |
+
dtype: float64
|
2423 |
+
- name: tokenmonster/englishcode-32000-consistent-v1
|
2424 |
+
dtype: float64
|
2425 |
+
- name: trimmed_cos_sim_to_canonical
|
2426 |
+
struct:
|
2427 |
+
- name: CohereLabs/aya-expanse-8b
|
2428 |
+
dtype: float64
|
2429 |
+
- name: Qwen/Qwen3-8B
|
2430 |
+
dtype: float64
|
2431 |
+
- name: bigscience/bloom
|
2432 |
+
dtype: float64
|
2433 |
+
- name: common-pile/comma-v0.1-1t
|
2434 |
+
dtype: float64
|
2435 |
+
- name: facebook/xglm-564M
|
2436 |
+
dtype: float64
|
2437 |
+
- name: google-bert/bert-base-multilingual-cased
|
2438 |
+
dtype: float64
|
2439 |
+
- name: google/byt5-small
|
2440 |
+
dtype: float64
|
2441 |
+
- name: google/gemma-2-2b
|
2442 |
+
dtype: float64
|
2443 |
+
- name: gpt2
|
2444 |
+
dtype: float64
|
2445 |
+
- name: meta-llama/Llama-3.2-1B
|
2446 |
+
dtype: float64
|
2447 |
+
- name: microsoft/Phi-3-mini-4k-instruct
|
2448 |
+
dtype: float64
|
2449 |
+
- name: mistralai/tekken
|
2450 |
+
dtype: float64
|
2451 |
+
- name: tiktoken/gpt-4o
|
2452 |
+
dtype: float64
|
2453 |
+
- name: tokenmonster/englishcode-32000-consistent-v1
|
2454 |
+
dtype: float64
|
2455 |
+
- name: token_counts
|
2456 |
+
struct:
|
2457 |
+
- name: CohereLabs/aya-expanse-8b
|
2458 |
+
dtype: int64
|
2459 |
+
- name: Qwen/Qwen3-8B
|
2460 |
+
dtype: int64
|
2461 |
+
- name: bigscience/bloom
|
2462 |
+
dtype: int64
|
2463 |
+
- name: common-pile/comma-v0.1-1t
|
2464 |
+
dtype: int64
|
2465 |
+
- name: facebook/xglm-564M
|
2466 |
+
dtype: int64
|
2467 |
+
- name: google-bert/bert-base-multilingual-cased
|
2468 |
+
dtype: int64
|
2469 |
+
- name: google/byt5-small
|
2470 |
+
dtype: int64
|
2471 |
+
- name: google/gemma-2-2b
|
2472 |
+
dtype: int64
|
2473 |
+
- name: gpt2
|
2474 |
+
dtype: int64
|
2475 |
+
- name: meta-llama/Llama-3.2-1B
|
2476 |
+
dtype: int64
|
2477 |
+
- name: microsoft/Phi-3-mini-4k-instruct
|
2478 |
+
dtype: int64
|
2479 |
+
- name: mistralai/tekken
|
2480 |
+
dtype: int64
|
2481 |
+
- name: tiktoken/gpt-4o
|
2482 |
+
dtype: int64
|
2483 |
+
- name: tokenmonster/englishcode-32000-consistent-v1
|
2484 |
+
dtype: int64
|
2485 |
+
splits:
|
2486 |
+
- name: test
|
2487 |
+
num_bytes: 22235
|
2488 |
+
num_examples: 42
|
2489 |
+
download_size: 40820
|
2490 |
+
dataset_size: 22235
|
2491 |
configs:
|
2492 |
- config_name: tokenizer_robustness_completion_english_abbreviations
|
2493 |
data_files:
|
|
|
2565 |
data_files:
|
2566 |
- split: test
|
2567 |
path: tokenizer_robustness_completion_english_macron_diacritic/test-*
|
2568 |
+
- config_name: tokenizer_robustness_completion_english_ocr_errors
|
2569 |
+
data_files:
|
2570 |
+
- split: test
|
2571 |
+
path: tokenizer_robustness_completion_english_ocr_errors/test-*
|
2572 |
---
|
2573 |
|
2574 |
# Dataset Card for Tokenization Robustness
|
tokenizer_robustness_completion_english_ocr_errors/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6dfe2aad92e89ab1d74090afe82f0814f0d72a437232e3220cdff4e74ba09137
|
3 |
+
size 40820
|