gsaltintas commited on
Commit
2fb20ba
·
verified ·
1 Parent(s): b737c5a

Uploading tokenizer_robustness_completion_stem_typographical_errors subset

Browse files
README.md CHANGED
@@ -1992,6 +1992,130 @@ dataset_info:
1992
  num_examples: 29
1993
  download_size: 32747
1994
  dataset_size: 17203
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1995
  configs:
1996
  - config_name: tokenizer_robustness_completion_stem_canonical
1997
  data_files:
@@ -2057,6 +2181,10 @@ configs:
2057
  data_files:
2058
  - split: test
2059
  path: tokenizer_robustness_completion_stem_superscript_subscript/test-*
 
 
 
 
2060
  ---
2061
 
2062
  # Dataset Card for Tokenization Robustness
 
1992
  num_examples: 29
1993
  download_size: 32747
1994
  dataset_size: 17203
1995
+ - config_name: tokenizer_robustness_completion_stem_typographical_errors
1996
+ features:
1997
+ - name: question
1998
+ dtype: string
1999
+ - name: choices
2000
+ list: string
2001
+ - name: answer
2002
+ dtype: int64
2003
+ - name: answer_label
2004
+ dtype: string
2005
+ - name: split
2006
+ dtype: string
2007
+ - name: subcategories
2008
+ dtype: string
2009
+ - name: lang
2010
+ dtype: string
2011
+ - name: second_lang
2012
+ dtype: string
2013
+ - name: notes
2014
+ dtype: string
2015
+ - name: id
2016
+ dtype: string
2017
+ - name: set_id
2018
+ dtype: string
2019
+ - name: variation_id
2020
+ dtype: string
2021
+ - name: question_general_category
2022
+ dtype: string
2023
+ - name: vanilla_cos_sim_to_canonical
2024
+ struct:
2025
+ - name: CohereLabs/aya-expanse-8b
2026
+ dtype: float64
2027
+ - name: Qwen/Qwen3-8B
2028
+ dtype: float64
2029
+ - name: bigscience/bloom
2030
+ dtype: float64
2031
+ - name: common-pile/comma-v0.1-1t
2032
+ dtype: float64
2033
+ - name: facebook/xglm-564M
2034
+ dtype: float64
2035
+ - name: google-bert/bert-base-multilingual-cased
2036
+ dtype: float64
2037
+ - name: google/byt5-small
2038
+ dtype: float64
2039
+ - name: google/gemma-2-2b
2040
+ dtype: float64
2041
+ - name: gpt2
2042
+ dtype: float64
2043
+ - name: meta-llama/Llama-3.2-1B
2044
+ dtype: float64
2045
+ - name: microsoft/Phi-3-mini-4k-instruct
2046
+ dtype: float64
2047
+ - name: mistralai/tekken
2048
+ dtype: float64
2049
+ - name: tiktoken/gpt-4o
2050
+ dtype: float64
2051
+ - name: tokenmonster/englishcode-32000-consistent-v1
2052
+ dtype: float64
2053
+ - name: trimmed_cos_sim_to_canonical
2054
+ struct:
2055
+ - name: CohereLabs/aya-expanse-8b
2056
+ dtype: float64
2057
+ - name: Qwen/Qwen3-8B
2058
+ dtype: float64
2059
+ - name: bigscience/bloom
2060
+ dtype: float64
2061
+ - name: common-pile/comma-v0.1-1t
2062
+ dtype: float64
2063
+ - name: facebook/xglm-564M
2064
+ dtype: float64
2065
+ - name: google-bert/bert-base-multilingual-cased
2066
+ dtype: float64
2067
+ - name: google/byt5-small
2068
+ dtype: float64
2069
+ - name: google/gemma-2-2b
2070
+ dtype: float64
2071
+ - name: gpt2
2072
+ dtype: float64
2073
+ - name: meta-llama/Llama-3.2-1B
2074
+ dtype: float64
2075
+ - name: microsoft/Phi-3-mini-4k-instruct
2076
+ dtype: float64
2077
+ - name: mistralai/tekken
2078
+ dtype: float64
2079
+ - name: tiktoken/gpt-4o
2080
+ dtype: float64
2081
+ - name: tokenmonster/englishcode-32000-consistent-v1
2082
+ dtype: float64
2083
+ - name: token_counts
2084
+ struct:
2085
+ - name: CohereLabs/aya-expanse-8b
2086
+ dtype: int64
2087
+ - name: Qwen/Qwen3-8B
2088
+ dtype: int64
2089
+ - name: bigscience/bloom
2090
+ dtype: int64
2091
+ - name: common-pile/comma-v0.1-1t
2092
+ dtype: int64
2093
+ - name: facebook/xglm-564M
2094
+ dtype: int64
2095
+ - name: google-bert/bert-base-multilingual-cased
2096
+ dtype: int64
2097
+ - name: google/byt5-small
2098
+ dtype: int64
2099
+ - name: google/gemma-2-2b
2100
+ dtype: int64
2101
+ - name: gpt2
2102
+ dtype: int64
2103
+ - name: meta-llama/Llama-3.2-1B
2104
+ dtype: int64
2105
+ - name: microsoft/Phi-3-mini-4k-instruct
2106
+ dtype: int64
2107
+ - name: mistralai/tekken
2108
+ dtype: int64
2109
+ - name: tiktoken/gpt-4o
2110
+ dtype: int64
2111
+ - name: tokenmonster/englishcode-32000-consistent-v1
2112
+ dtype: int64
2113
+ splits:
2114
+ - name: test
2115
+ num_bytes: 21308
2116
+ num_examples: 38
2117
+ download_size: 39429
2118
+ dataset_size: 21308
2119
  configs:
2120
  - config_name: tokenizer_robustness_completion_stem_canonical
2121
  data_files:
 
2181
  data_files:
2182
  - split: test
2183
  path: tokenizer_robustness_completion_stem_superscript_subscript/test-*
2184
+ - config_name: tokenizer_robustness_completion_stem_typographical_errors
2185
+ data_files:
2186
+ - split: test
2187
+ path: tokenizer_robustness_completion_stem_typographical_errors/test-*
2188
  ---
2189
 
2190
  # Dataset Card for Tokenization Robustness
tokenizer_robustness_completion_stem_typographical_errors/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c34bd94372bd3d1d774b4c33e878a32da12717a782e216a7371af302e7f1371
3
+ size 39429