Uploading tokenizer_robustness_completion_english_lowercase subset
Browse files
README.md
CHANGED
@@ -2116,6 +2116,130 @@ dataset_info:
|
|
2116 |
num_examples: 41
|
2117 |
download_size: 40349
|
2118 |
dataset_size: 23047
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2119 |
configs:
|
2120 |
- config_name: tokenizer_robustness_completion_english_abbreviations
|
2121 |
data_files:
|
@@ -2185,6 +2309,10 @@ configs:
|
|
2185 |
data_files:
|
2186 |
- split: test
|
2187 |
path: tokenizer_robustness_completion_english_letter_repetition_for_emphasis/test-*
|
|
|
|
|
|
|
|
|
2188 |
---
|
2189 |
|
2190 |
# Dataset Card for Tokenization Robustness
|
|
|
2116 |
num_examples: 41
|
2117 |
download_size: 40349
|
2118 |
dataset_size: 23047
|
2119 |
+
- config_name: tokenizer_robustness_completion_english_lowercase
|
2120 |
+
features:
|
2121 |
+
- name: question
|
2122 |
+
dtype: string
|
2123 |
+
- name: choices
|
2124 |
+
list: string
|
2125 |
+
- name: answer
|
2126 |
+
dtype: int64
|
2127 |
+
- name: answer_label
|
2128 |
+
dtype: string
|
2129 |
+
- name: split
|
2130 |
+
dtype: string
|
2131 |
+
- name: subcategories
|
2132 |
+
dtype: string
|
2133 |
+
- name: category
|
2134 |
+
dtype: string
|
2135 |
+
- name: lang
|
2136 |
+
dtype: string
|
2137 |
+
- name: second_lang
|
2138 |
+
dtype: string
|
2139 |
+
- name: notes
|
2140 |
+
dtype: string
|
2141 |
+
- name: id
|
2142 |
+
dtype: string
|
2143 |
+
- name: set_id
|
2144 |
+
dtype: string
|
2145 |
+
- name: variation_id
|
2146 |
+
dtype: string
|
2147 |
+
- name: vanilla_cos_sim_to_canonical
|
2148 |
+
struct:
|
2149 |
+
- name: CohereLabs/aya-expanse-8b
|
2150 |
+
dtype: float64
|
2151 |
+
- name: Qwen/Qwen3-8B
|
2152 |
+
dtype: float64
|
2153 |
+
- name: bigscience/bloom
|
2154 |
+
dtype: float64
|
2155 |
+
- name: common-pile/comma-v0.1-1t
|
2156 |
+
dtype: float64
|
2157 |
+
- name: facebook/xglm-564M
|
2158 |
+
dtype: float64
|
2159 |
+
- name: google-bert/bert-base-multilingual-cased
|
2160 |
+
dtype: float64
|
2161 |
+
- name: google/byt5-small
|
2162 |
+
dtype: float64
|
2163 |
+
- name: google/gemma-2-2b
|
2164 |
+
dtype: float64
|
2165 |
+
- name: gpt2
|
2166 |
+
dtype: float64
|
2167 |
+
- name: meta-llama/Llama-3.2-1B
|
2168 |
+
dtype: float64
|
2169 |
+
- name: microsoft/Phi-3-mini-4k-instruct
|
2170 |
+
dtype: float64
|
2171 |
+
- name: mistralai/tekken
|
2172 |
+
dtype: float64
|
2173 |
+
- name: tiktoken/gpt-4o
|
2174 |
+
dtype: float64
|
2175 |
+
- name: tokenmonster/englishcode-32000-consistent-v1
|
2176 |
+
dtype: float64
|
2177 |
+
- name: trimmed_cos_sim_to_canonical
|
2178 |
+
struct:
|
2179 |
+
- name: CohereLabs/aya-expanse-8b
|
2180 |
+
dtype: float64
|
2181 |
+
- name: Qwen/Qwen3-8B
|
2182 |
+
dtype: float64
|
2183 |
+
- name: bigscience/bloom
|
2184 |
+
dtype: float64
|
2185 |
+
- name: common-pile/comma-v0.1-1t
|
2186 |
+
dtype: float64
|
2187 |
+
- name: facebook/xglm-564M
|
2188 |
+
dtype: float64
|
2189 |
+
- name: google-bert/bert-base-multilingual-cased
|
2190 |
+
dtype: float64
|
2191 |
+
- name: google/byt5-small
|
2192 |
+
dtype: float64
|
2193 |
+
- name: google/gemma-2-2b
|
2194 |
+
dtype: float64
|
2195 |
+
- name: gpt2
|
2196 |
+
dtype: float64
|
2197 |
+
- name: meta-llama/Llama-3.2-1B
|
2198 |
+
dtype: float64
|
2199 |
+
- name: microsoft/Phi-3-mini-4k-instruct
|
2200 |
+
dtype: float64
|
2201 |
+
- name: mistralai/tekken
|
2202 |
+
dtype: float64
|
2203 |
+
- name: tiktoken/gpt-4o
|
2204 |
+
dtype: float64
|
2205 |
+
- name: tokenmonster/englishcode-32000-consistent-v1
|
2206 |
+
dtype: float64
|
2207 |
+
- name: token_counts
|
2208 |
+
struct:
|
2209 |
+
- name: CohereLabs/aya-expanse-8b
|
2210 |
+
dtype: int64
|
2211 |
+
- name: Qwen/Qwen3-8B
|
2212 |
+
dtype: int64
|
2213 |
+
- name: bigscience/bloom
|
2214 |
+
dtype: int64
|
2215 |
+
- name: common-pile/comma-v0.1-1t
|
2216 |
+
dtype: int64
|
2217 |
+
- name: facebook/xglm-564M
|
2218 |
+
dtype: int64
|
2219 |
+
- name: google-bert/bert-base-multilingual-cased
|
2220 |
+
dtype: int64
|
2221 |
+
- name: google/byt5-small
|
2222 |
+
dtype: int64
|
2223 |
+
- name: google/gemma-2-2b
|
2224 |
+
dtype: int64
|
2225 |
+
- name: gpt2
|
2226 |
+
dtype: int64
|
2227 |
+
- name: meta-llama/Llama-3.2-1B
|
2228 |
+
dtype: int64
|
2229 |
+
- name: microsoft/Phi-3-mini-4k-instruct
|
2230 |
+
dtype: int64
|
2231 |
+
- name: mistralai/tekken
|
2232 |
+
dtype: int64
|
2233 |
+
- name: tiktoken/gpt-4o
|
2234 |
+
dtype: int64
|
2235 |
+
- name: tokenmonster/englishcode-32000-consistent-v1
|
2236 |
+
dtype: int64
|
2237 |
+
splits:
|
2238 |
+
- name: test
|
2239 |
+
num_bytes: 20307
|
2240 |
+
num_examples: 38
|
2241 |
+
download_size: 37855
|
2242 |
+
dataset_size: 20307
|
2243 |
configs:
|
2244 |
- config_name: tokenizer_robustness_completion_english_abbreviations
|
2245 |
data_files:
|
|
|
2309 |
data_files:
|
2310 |
- split: test
|
2311 |
path: tokenizer_robustness_completion_english_letter_repetition_for_emphasis/test-*
|
2312 |
+
- config_name: tokenizer_robustness_completion_english_lowercase
|
2313 |
+
data_files:
|
2314 |
+
- split: test
|
2315 |
+
path: tokenizer_robustness_completion_english_lowercase/test-*
|
2316 |
---
|
2317 |
|
2318 |
# Dataset Card for Tokenization Robustness
|
tokenizer_robustness_completion_english_lowercase/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8996318a115c3b3f5c1b38b173e27d4012fdd8182faa1e21bb09d10222ff7a54
|
3 |
+
size 37855
|