gsaltintas commited on
Commit
4ae7467
·
verified ·
1 Parent(s): 496728b

Uploading tokenizer_robustness_completion_english_character_substitution subset

Browse files
README.md CHANGED
@@ -504,6 +504,130 @@ dataset_info:
504
  num_examples: 40
505
  download_size: 40840
506
  dataset_size: 22322
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
507
  configs:
508
  - config_name: tokenizer_robustness_completion_english_abbreviations
509
  data_files:
@@ -521,6 +645,10 @@ configs:
521
  data_files:
522
  - split: test
523
  path: tokenizer_robustness_completion_english_character_deletion/test-*
 
 
 
 
524
  ---
525
 
526
  # Dataset Card for Tokenization Robustness
 
504
  num_examples: 40
505
  download_size: 40840
506
  dataset_size: 22322
507
+ - config_name: tokenizer_robustness_completion_english_character_substitution
508
+ features:
509
+ - name: question
510
+ dtype: string
511
+ - name: choices
512
+ list: string
513
+ - name: answer
514
+ dtype: int64
515
+ - name: answer_label
516
+ dtype: string
517
+ - name: split
518
+ dtype: string
519
+ - name: subcategories
520
+ dtype: string
521
+ - name: category
522
+ dtype: string
523
+ - name: lang
524
+ dtype: string
525
+ - name: second_lang
526
+ dtype: string
527
+ - name: notes
528
+ dtype: string
529
+ - name: id
530
+ dtype: string
531
+ - name: set_id
532
+ dtype: string
533
+ - name: variation_id
534
+ dtype: string
535
+ - name: vanilla_cos_sim_to_canonical
536
+ struct:
537
+ - name: CohereLabs/aya-expanse-8b
538
+ dtype: float64
539
+ - name: Qwen/Qwen3-8B
540
+ dtype: float64
541
+ - name: bigscience/bloom
542
+ dtype: float64
543
+ - name: common-pile/comma-v0.1-1t
544
+ dtype: float64
545
+ - name: facebook/xglm-564M
546
+ dtype: float64
547
+ - name: google-bert/bert-base-multilingual-cased
548
+ dtype: float64
549
+ - name: google/byt5-small
550
+ dtype: float64
551
+ - name: google/gemma-2-2b
552
+ dtype: float64
553
+ - name: gpt2
554
+ dtype: float64
555
+ - name: meta-llama/Llama-3.2-1B
556
+ dtype: float64
557
+ - name: microsoft/Phi-3-mini-4k-instruct
558
+ dtype: float64
559
+ - name: mistralai/tekken
560
+ dtype: float64
561
+ - name: tiktoken/gpt-4o
562
+ dtype: float64
563
+ - name: tokenmonster/englishcode-32000-consistent-v1
564
+ dtype: float64
565
+ - name: trimmed_cos_sim_to_canonical
566
+ struct:
567
+ - name: CohereLabs/aya-expanse-8b
568
+ dtype: float64
569
+ - name: Qwen/Qwen3-8B
570
+ dtype: float64
571
+ - name: bigscience/bloom
572
+ dtype: float64
573
+ - name: common-pile/comma-v0.1-1t
574
+ dtype: float64
575
+ - name: facebook/xglm-564M
576
+ dtype: float64
577
+ - name: google-bert/bert-base-multilingual-cased
578
+ dtype: float64
579
+ - name: google/byt5-small
580
+ dtype: float64
581
+ - name: google/gemma-2-2b
582
+ dtype: float64
583
+ - name: gpt2
584
+ dtype: float64
585
+ - name: meta-llama/Llama-3.2-1B
586
+ dtype: float64
587
+ - name: microsoft/Phi-3-mini-4k-instruct
588
+ dtype: float64
589
+ - name: mistralai/tekken
590
+ dtype: float64
591
+ - name: tiktoken/gpt-4o
592
+ dtype: float64
593
+ - name: tokenmonster/englishcode-32000-consistent-v1
594
+ dtype: float64
595
+ - name: token_counts
596
+ struct:
597
+ - name: CohereLabs/aya-expanse-8b
598
+ dtype: int64
599
+ - name: Qwen/Qwen3-8B
600
+ dtype: int64
601
+ - name: bigscience/bloom
602
+ dtype: int64
603
+ - name: common-pile/comma-v0.1-1t
604
+ dtype: int64
605
+ - name: facebook/xglm-564M
606
+ dtype: int64
607
+ - name: google-bert/bert-base-multilingual-cased
608
+ dtype: int64
609
+ - name: google/byt5-small
610
+ dtype: int64
611
+ - name: google/gemma-2-2b
612
+ dtype: int64
613
+ - name: gpt2
614
+ dtype: int64
615
+ - name: meta-llama/Llama-3.2-1B
616
+ dtype: int64
617
+ - name: microsoft/Phi-3-mini-4k-instruct
618
+ dtype: int64
619
+ - name: mistralai/tekken
620
+ dtype: int64
621
+ - name: tiktoken/gpt-4o
622
+ dtype: int64
623
+ - name: tokenmonster/englishcode-32000-consistent-v1
624
+ dtype: int64
625
+ splits:
626
+ - name: test
627
+ num_bytes: 8942
628
+ num_examples: 16
629
+ download_size: 32517
630
+ dataset_size: 8942
631
  configs:
632
  - config_name: tokenizer_robustness_completion_english_abbreviations
633
  data_files:
 
645
  data_files:
646
  - split: test
647
  path: tokenizer_robustness_completion_english_character_deletion/test-*
648
+ - config_name: tokenizer_robustness_completion_english_character_substitution
649
+ data_files:
650
+ - split: test
651
+ path: tokenizer_robustness_completion_english_character_substitution/test-*
652
  ---
653
 
654
  # Dataset Card for Tokenization Robustness
tokenizer_robustness_completion_english_character_substitution/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:714ea34f506eb05e98ff38b9abdd0e7f7261e32c532005221f7cb46f5f932fe9
3
+ size 32517