Datasets:
Uploading tokenizer_robustness_completion_english_canonical subset
Browse files
README.md
CHANGED
@@ -8,7 +8,7 @@ tags:
|
|
8 |
- multilingual
|
9 |
- tokenization
|
10 |
dataset_info:
|
11 |
-
|
12 |
features:
|
13 |
- name: question
|
14 |
dtype: string
|
@@ -132,11 +132,139 @@ dataset_info:
|
|
132 |
num_examples: 37
|
133 |
download_size: 36121
|
134 |
dataset_size: 19544
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
135 |
configs:
|
136 |
- config_name: tokenizer_robustness_completion_english_abbreviations
|
137 |
data_files:
|
138 |
- split: test
|
139 |
path: tokenizer_robustness_completion_english_abbreviations/test-*
|
|
|
|
|
|
|
|
|
140 |
---
|
141 |
|
142 |
# Dataset Card for Tokenization Robustness
|
|
|
8 |
- multilingual
|
9 |
- tokenization
|
10 |
dataset_info:
|
11 |
+
- config_name: tokenizer_robustness_completion_english_abbreviations
|
12 |
features:
|
13 |
- name: question
|
14 |
dtype: string
|
|
|
132 |
num_examples: 37
|
133 |
download_size: 36121
|
134 |
dataset_size: 19544
|
135 |
+
- config_name: tokenizer_robustness_completion_english_canonical
|
136 |
+
features:
|
137 |
+
- name: question
|
138 |
+
dtype: string
|
139 |
+
- name: choices
|
140 |
+
list: string
|
141 |
+
- name: answer
|
142 |
+
dtype: int64
|
143 |
+
- name: answer_label
|
144 |
+
dtype: string
|
145 |
+
- name: split
|
146 |
+
dtype: string
|
147 |
+
- name: subcategories
|
148 |
+
dtype: string
|
149 |
+
- name: category
|
150 |
+
dtype: string
|
151 |
+
- name: lang
|
152 |
+
dtype: string
|
153 |
+
- name: second_lang
|
154 |
+
dtype: string
|
155 |
+
- name: notes
|
156 |
+
dtype: string
|
157 |
+
- name: id
|
158 |
+
dtype: string
|
159 |
+
- name: set_id
|
160 |
+
dtype: string
|
161 |
+
- name: variation_id
|
162 |
+
dtype: string
|
163 |
+
- name: vanilla_cos_sim_to_canonical
|
164 |
+
struct:
|
165 |
+
- name: CohereLabs/aya-expanse-8b
|
166 |
+
dtype: float64
|
167 |
+
- name: Qwen/Qwen3-8B
|
168 |
+
dtype: float64
|
169 |
+
- name: bigscience/bloom
|
170 |
+
dtype: float64
|
171 |
+
- name: common-pile/comma-v0.1-1t
|
172 |
+
dtype: float64
|
173 |
+
- name: facebook/xglm-564M
|
174 |
+
dtype: float64
|
175 |
+
- name: google-bert/bert-base-multilingual-cased
|
176 |
+
dtype: float64
|
177 |
+
- name: google/byt5-small
|
178 |
+
dtype: float64
|
179 |
+
- name: google/gemma-2-2b
|
180 |
+
dtype: float64
|
181 |
+
- name: gpt2
|
182 |
+
dtype: float64
|
183 |
+
- name: meta-llama/Llama-3.2-1B
|
184 |
+
dtype: float64
|
185 |
+
- name: microsoft/Phi-3-mini-4k-instruct
|
186 |
+
dtype: float64
|
187 |
+
- name: mistralai/tekken
|
188 |
+
dtype: float64
|
189 |
+
- name: tiktoken/gpt-4o
|
190 |
+
dtype: float64
|
191 |
+
- name: tokenmonster/englishcode-32000-consistent-v1
|
192 |
+
dtype: float64
|
193 |
+
- name: trimmed_cos_sim_to_canonical
|
194 |
+
struct:
|
195 |
+
- name: CohereLabs/aya-expanse-8b
|
196 |
+
dtype: float64
|
197 |
+
- name: Qwen/Qwen3-8B
|
198 |
+
dtype: float64
|
199 |
+
- name: bigscience/bloom
|
200 |
+
dtype: float64
|
201 |
+
- name: common-pile/comma-v0.1-1t
|
202 |
+
dtype: float64
|
203 |
+
- name: facebook/xglm-564M
|
204 |
+
dtype: float64
|
205 |
+
- name: google-bert/bert-base-multilingual-cased
|
206 |
+
dtype: float64
|
207 |
+
- name: google/byt5-small
|
208 |
+
dtype: float64
|
209 |
+
- name: google/gemma-2-2b
|
210 |
+
dtype: float64
|
211 |
+
- name: gpt2
|
212 |
+
dtype: float64
|
213 |
+
- name: meta-llama/Llama-3.2-1B
|
214 |
+
dtype: float64
|
215 |
+
- name: microsoft/Phi-3-mini-4k-instruct
|
216 |
+
dtype: float64
|
217 |
+
- name: mistralai/tekken
|
218 |
+
dtype: float64
|
219 |
+
- name: tiktoken/gpt-4o
|
220 |
+
dtype: float64
|
221 |
+
- name: tokenmonster/englishcode-32000-consistent-v1
|
222 |
+
dtype: float64
|
223 |
+
- name: token_counts
|
224 |
+
struct:
|
225 |
+
- name: CohereLabs/aya-expanse-8b
|
226 |
+
dtype: int64
|
227 |
+
- name: Qwen/Qwen3-8B
|
228 |
+
dtype: int64
|
229 |
+
- name: bigscience/bloom
|
230 |
+
dtype: int64
|
231 |
+
- name: common-pile/comma-v0.1-1t
|
232 |
+
dtype: int64
|
233 |
+
- name: facebook/xglm-564M
|
234 |
+
dtype: int64
|
235 |
+
- name: google-bert/bert-base-multilingual-cased
|
236 |
+
dtype: int64
|
237 |
+
- name: google/byt5-small
|
238 |
+
dtype: int64
|
239 |
+
- name: google/gemma-2-2b
|
240 |
+
dtype: int64
|
241 |
+
- name: gpt2
|
242 |
+
dtype: int64
|
243 |
+
- name: meta-llama/Llama-3.2-1B
|
244 |
+
dtype: int64
|
245 |
+
- name: microsoft/Phi-3-mini-4k-instruct
|
246 |
+
dtype: int64
|
247 |
+
- name: mistralai/tekken
|
248 |
+
dtype: int64
|
249 |
+
- name: tiktoken/gpt-4o
|
250 |
+
dtype: int64
|
251 |
+
- name: tokenmonster/englishcode-32000-consistent-v1
|
252 |
+
dtype: int64
|
253 |
+
splits:
|
254 |
+
- name: test
|
255 |
+
num_bytes: 21874
|
256 |
+
num_examples: 40
|
257 |
+
download_size: 32721
|
258 |
+
dataset_size: 21874
|
259 |
configs:
|
260 |
- config_name: tokenizer_robustness_completion_english_abbreviations
|
261 |
data_files:
|
262 |
- split: test
|
263 |
path: tokenizer_robustness_completion_english_abbreviations/test-*
|
264 |
+
- config_name: tokenizer_robustness_completion_english_canonical
|
265 |
+
data_files:
|
266 |
+
- split: test
|
267 |
+
path: tokenizer_robustness_completion_english_canonical/test-*
|
268 |
---
|
269 |
|
270 |
# Dataset Card for Tokenization Robustness
|
tokenizer_robustness_completion_english_canonical/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:58c75b0aedfdc3250649d9afa938fde6ccda4e681fbff49ee8720e268ea08149
|
3 |
+
size 32721
|