gsaltintas commited on
Commit
74e7065
·
verified ·
1 Parent(s): 9a05055

Uploading tokenizer_robustness_completion_stem_equivalent_expressions subset

Browse files
README.md CHANGED
@@ -876,6 +876,130 @@ dataset_info:
876
  num_examples: 42
877
  download_size: 32767
878
  dataset_size: 25460
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
879
  configs:
880
  - config_name: tokenizer_robustness_completion_stem_canonical
881
  data_files:
@@ -905,6 +1029,10 @@ configs:
905
  data_files:
906
  - split: test
907
  path: tokenizer_robustness_completion_stem_enclosed_characters/test-*
 
 
 
 
908
  ---
909
 
910
  # Dataset Card for Tokenization Robustness
 
876
  num_examples: 42
877
  download_size: 32767
878
  dataset_size: 25460
879
+ - config_name: tokenizer_robustness_completion_stem_equivalent_expressions
880
+ features:
881
+ - name: question
882
+ dtype: string
883
+ - name: choices
884
+ list: string
885
+ - name: answer
886
+ dtype: int64
887
+ - name: answer_label
888
+ dtype: string
889
+ - name: split
890
+ dtype: string
891
+ - name: subcategories
892
+ dtype: string
893
+ - name: lang
894
+ dtype: string
895
+ - name: second_lang
896
+ dtype: string
897
+ - name: notes
898
+ dtype: string
899
+ - name: id
900
+ dtype: string
901
+ - name: set_id
902
+ dtype: string
903
+ - name: variation_id
904
+ dtype: string
905
+ - name: question_general_category
906
+ dtype: string
907
+ - name: vanilla_cos_sim_to_canonical
908
+ struct:
909
+ - name: CohereLabs/aya-expanse-8b
910
+ dtype: float64
911
+ - name: Qwen/Qwen3-8B
912
+ dtype: float64
913
+ - name: bigscience/bloom
914
+ dtype: float64
915
+ - name: common-pile/comma-v0.1-1t
916
+ dtype: float64
917
+ - name: facebook/xglm-564M
918
+ dtype: float64
919
+ - name: google-bert/bert-base-multilingual-cased
920
+ dtype: float64
921
+ - name: google/byt5-small
922
+ dtype: float64
923
+ - name: google/gemma-2-2b
924
+ dtype: float64
925
+ - name: gpt2
926
+ dtype: float64
927
+ - name: meta-llama/Llama-3.2-1B
928
+ dtype: float64
929
+ - name: microsoft/Phi-3-mini-4k-instruct
930
+ dtype: float64
931
+ - name: mistralai/tekken
932
+ dtype: float64
933
+ - name: tiktoken/gpt-4o
934
+ dtype: float64
935
+ - name: tokenmonster/englishcode-32000-consistent-v1
936
+ dtype: float64
937
+ - name: trimmed_cos_sim_to_canonical
938
+ struct:
939
+ - name: CohereLabs/aya-expanse-8b
940
+ dtype: float64
941
+ - name: Qwen/Qwen3-8B
942
+ dtype: float64
943
+ - name: bigscience/bloom
944
+ dtype: float64
945
+ - name: common-pile/comma-v0.1-1t
946
+ dtype: float64
947
+ - name: facebook/xglm-564M
948
+ dtype: float64
949
+ - name: google-bert/bert-base-multilingual-cased
950
+ dtype: float64
951
+ - name: google/byt5-small
952
+ dtype: float64
953
+ - name: google/gemma-2-2b
954
+ dtype: float64
955
+ - name: gpt2
956
+ dtype: float64
957
+ - name: meta-llama/Llama-3.2-1B
958
+ dtype: float64
959
+ - name: microsoft/Phi-3-mini-4k-instruct
960
+ dtype: float64
961
+ - name: mistralai/tekken
962
+ dtype: float64
963
+ - name: tiktoken/gpt-4o
964
+ dtype: float64
965
+ - name: tokenmonster/englishcode-32000-consistent-v1
966
+ dtype: float64
967
+ - name: token_counts
968
+ struct:
969
+ - name: CohereLabs/aya-expanse-8b
970
+ dtype: int64
971
+ - name: Qwen/Qwen3-8B
972
+ dtype: int64
973
+ - name: bigscience/bloom
974
+ dtype: int64
975
+ - name: common-pile/comma-v0.1-1t
976
+ dtype: int64
977
+ - name: facebook/xglm-564M
978
+ dtype: int64
979
+ - name: google-bert/bert-base-multilingual-cased
980
+ dtype: int64
981
+ - name: google/byt5-small
982
+ dtype: int64
983
+ - name: google/gemma-2-2b
984
+ dtype: int64
985
+ - name: gpt2
986
+ dtype: int64
987
+ - name: meta-llama/Llama-3.2-1B
988
+ dtype: int64
989
+ - name: microsoft/Phi-3-mini-4k-instruct
990
+ dtype: int64
991
+ - name: mistralai/tekken
992
+ dtype: int64
993
+ - name: tiktoken/gpt-4o
994
+ dtype: int64
995
+ - name: tokenmonster/englishcode-32000-consistent-v1
996
+ dtype: int64
997
+ splits:
998
+ - name: test
999
+ num_bytes: 8286
1000
+ num_examples: 15
1001
+ download_size: 32741
1002
+ dataset_size: 8286
1003
  configs:
1004
  - config_name: tokenizer_robustness_completion_stem_canonical
1005
  data_files:
 
1029
  data_files:
1030
  - split: test
1031
  path: tokenizer_robustness_completion_stem_enclosed_characters/test-*
1032
+ - config_name: tokenizer_robustness_completion_stem_equivalent_expressions
1033
+ data_files:
1034
+ - split: test
1035
+ path: tokenizer_robustness_completion_stem_equivalent_expressions/test-*
1036
  ---
1037
 
1038
  # Dataset Card for Tokenization Robustness
tokenizer_robustness_completion_stem_equivalent_expressions/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b561549845e0015175c025ba02cd9289a141ec462a667310d4cbd3d618cdb68
3
+ size 32741