gsaltintas commited on
Commit
5b02514
·
verified ·
1 Parent(s): 951f7ca

Uploading tokenizer_robustness_completion_english_capitalization subset

Browse files
README.md CHANGED
@@ -256,6 +256,130 @@ dataset_info:
256
  num_examples: 40
257
  download_size: 32721
258
  dataset_size: 21874
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
259
  configs:
260
  - config_name: tokenizer_robustness_completion_english_abbreviations
261
  data_files:
@@ -265,6 +389,10 @@ configs:
265
  data_files:
266
  - split: test
267
  path: tokenizer_robustness_completion_english_canonical/test-*
 
 
 
 
268
  ---
269
 
270
  # Dataset Card for Tokenization Robustness
 
256
  num_examples: 40
257
  download_size: 32721
258
  dataset_size: 21874
259
+ - config_name: tokenizer_robustness_completion_english_capitalization
260
+ features:
261
+ - name: question
262
+ dtype: string
263
+ - name: choices
264
+ list: string
265
+ - name: answer
266
+ dtype: int64
267
+ - name: answer_label
268
+ dtype: string
269
+ - name: split
270
+ dtype: string
271
+ - name: subcategories
272
+ dtype: string
273
+ - name: category
274
+ dtype: string
275
+ - name: lang
276
+ dtype: string
277
+ - name: second_lang
278
+ dtype: string
279
+ - name: notes
280
+ dtype: string
281
+ - name: id
282
+ dtype: string
283
+ - name: set_id
284
+ dtype: string
285
+ - name: variation_id
286
+ dtype: string
287
+ - name: vanilla_cos_sim_to_canonical
288
+ struct:
289
+ - name: CohereLabs/aya-expanse-8b
290
+ dtype: float64
291
+ - name: Qwen/Qwen3-8B
292
+ dtype: float64
293
+ - name: bigscience/bloom
294
+ dtype: float64
295
+ - name: common-pile/comma-v0.1-1t
296
+ dtype: float64
297
+ - name: facebook/xglm-564M
298
+ dtype: float64
299
+ - name: google-bert/bert-base-multilingual-cased
300
+ dtype: float64
301
+ - name: google/byt5-small
302
+ dtype: float64
303
+ - name: google/gemma-2-2b
304
+ dtype: float64
305
+ - name: gpt2
306
+ dtype: float64
307
+ - name: meta-llama/Llama-3.2-1B
308
+ dtype: float64
309
+ - name: microsoft/Phi-3-mini-4k-instruct
310
+ dtype: float64
311
+ - name: mistralai/tekken
312
+ dtype: float64
313
+ - name: tiktoken/gpt-4o
314
+ dtype: float64
315
+ - name: tokenmonster/englishcode-32000-consistent-v1
316
+ dtype: float64
317
+ - name: trimmed_cos_sim_to_canonical
318
+ struct:
319
+ - name: CohereLabs/aya-expanse-8b
320
+ dtype: float64
321
+ - name: Qwen/Qwen3-8B
322
+ dtype: float64
323
+ - name: bigscience/bloom
324
+ dtype: float64
325
+ - name: common-pile/comma-v0.1-1t
326
+ dtype: float64
327
+ - name: facebook/xglm-564M
328
+ dtype: float64
329
+ - name: google-bert/bert-base-multilingual-cased
330
+ dtype: float64
331
+ - name: google/byt5-small
332
+ dtype: float64
333
+ - name: google/gemma-2-2b
334
+ dtype: float64
335
+ - name: gpt2
336
+ dtype: float64
337
+ - name: meta-llama/Llama-3.2-1B
338
+ dtype: float64
339
+ - name: microsoft/Phi-3-mini-4k-instruct
340
+ dtype: float64
341
+ - name: mistralai/tekken
342
+ dtype: float64
343
+ - name: tiktoken/gpt-4o
344
+ dtype: float64
345
+ - name: tokenmonster/englishcode-32000-consistent-v1
346
+ dtype: float64
347
+ - name: token_counts
348
+ struct:
349
+ - name: CohereLabs/aya-expanse-8b
350
+ dtype: int64
351
+ - name: Qwen/Qwen3-8B
352
+ dtype: int64
353
+ - name: bigscience/bloom
354
+ dtype: int64
355
+ - name: common-pile/comma-v0.1-1t
356
+ dtype: int64
357
+ - name: facebook/xglm-564M
358
+ dtype: int64
359
+ - name: google-bert/bert-base-multilingual-cased
360
+ dtype: int64
361
+ - name: google/byt5-small
362
+ dtype: int64
363
+ - name: google/gemma-2-2b
364
+ dtype: int64
365
+ - name: gpt2
366
+ dtype: int64
367
+ - name: meta-llama/Llama-3.2-1B
368
+ dtype: int64
369
+ - name: microsoft/Phi-3-mini-4k-instruct
370
+ dtype: int64
371
+ - name: mistralai/tekken
372
+ dtype: int64
373
+ - name: tiktoken/gpt-4o
374
+ dtype: int64
375
+ - name: tokenmonster/englishcode-32000-consistent-v1
376
+ dtype: int64
377
+ splits:
378
+ - name: test
379
+ num_bytes: 21413
380
+ num_examples: 40
381
+ download_size: 40005
382
+ dataset_size: 21413
383
  configs:
384
  - config_name: tokenizer_robustness_completion_english_abbreviations
385
  data_files:
 
389
  data_files:
390
  - split: test
391
  path: tokenizer_robustness_completion_english_canonical/test-*
392
+ - config_name: tokenizer_robustness_completion_english_capitalization
393
+ data_files:
394
+ - split: test
395
+ path: tokenizer_robustness_completion_english_capitalization/test-*
396
  ---
397
 
398
  # Dataset Card for Tokenization Robustness
tokenizer_robustness_completion_english_capitalization/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45c30c5a021d5f0d98165a8a9ced89e0e69a671dc7ad740f42e4bb6a7221cf08
3
+ size 40005