gsaltintas commited on
Commit
ba82fca
·
verified ·
1 Parent(s): 85817bc

Uploading tokenizer_robustness_completion_stem_compounds subset

Browse files
README.md CHANGED
@@ -380,6 +380,130 @@ dataset_info:
380
  num_examples: 15
381
  download_size: 32313
382
  dataset_size: 7820
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
383
  configs:
384
  - config_name: tokenizer_robustness_completion_stem_canonical
385
  data_files:
@@ -393,6 +517,10 @@ configs:
393
  data_files:
394
  - split: test
395
  path: tokenizer_robustness_completion_stem_colloquial/test-*
 
 
 
 
396
  ---
397
 
398
  # Dataset Card for Tokenization Robustness
 
380
  num_examples: 15
381
  download_size: 32313
382
  dataset_size: 7820
383
+ - config_name: tokenizer_robustness_completion_stem_compounds
384
+ features:
385
+ - name: question
386
+ dtype: string
387
+ - name: choices
388
+ list: string
389
+ - name: answer
390
+ dtype: int64
391
+ - name: answer_label
392
+ dtype: string
393
+ - name: split
394
+ dtype: string
395
+ - name: subcategories
396
+ dtype: string
397
+ - name: lang
398
+ dtype: string
399
+ - name: second_lang
400
+ dtype: string
401
+ - name: notes
402
+ dtype: string
403
+ - name: id
404
+ dtype: string
405
+ - name: set_id
406
+ dtype: string
407
+ - name: variation_id
408
+ dtype: string
409
+ - name: question_general_category
410
+ dtype: string
411
+ - name: vanilla_cos_sim_to_canonical
412
+ struct:
413
+ - name: CohereLabs/aya-expanse-8b
414
+ dtype: float64
415
+ - name: Qwen/Qwen3-8B
416
+ dtype: float64
417
+ - name: bigscience/bloom
418
+ dtype: float64
419
+ - name: common-pile/comma-v0.1-1t
420
+ dtype: float64
421
+ - name: facebook/xglm-564M
422
+ dtype: float64
423
+ - name: google-bert/bert-base-multilingual-cased
424
+ dtype: float64
425
+ - name: google/byt5-small
426
+ dtype: float64
427
+ - name: google/gemma-2-2b
428
+ dtype: float64
429
+ - name: gpt2
430
+ dtype: float64
431
+ - name: meta-llama/Llama-3.2-1B
432
+ dtype: float64
433
+ - name: microsoft/Phi-3-mini-4k-instruct
434
+ dtype: float64
435
+ - name: mistralai/tekken
436
+ dtype: float64
437
+ - name: tiktoken/gpt-4o
438
+ dtype: float64
439
+ - name: tokenmonster/englishcode-32000-consistent-v1
440
+ dtype: float64
441
+ - name: trimmed_cos_sim_to_canonical
442
+ struct:
443
+ - name: CohereLabs/aya-expanse-8b
444
+ dtype: float64
445
+ - name: Qwen/Qwen3-8B
446
+ dtype: float64
447
+ - name: bigscience/bloom
448
+ dtype: float64
449
+ - name: common-pile/comma-v0.1-1t
450
+ dtype: float64
451
+ - name: facebook/xglm-564M
452
+ dtype: float64
453
+ - name: google-bert/bert-base-multilingual-cased
454
+ dtype: float64
455
+ - name: google/byt5-small
456
+ dtype: float64
457
+ - name: google/gemma-2-2b
458
+ dtype: float64
459
+ - name: gpt2
460
+ dtype: float64
461
+ - name: meta-llama/Llama-3.2-1B
462
+ dtype: float64
463
+ - name: microsoft/Phi-3-mini-4k-instruct
464
+ dtype: float64
465
+ - name: mistralai/tekken
466
+ dtype: float64
467
+ - name: tiktoken/gpt-4o
468
+ dtype: float64
469
+ - name: tokenmonster/englishcode-32000-consistent-v1
470
+ dtype: float64
471
+ - name: token_counts
472
+ struct:
473
+ - name: CohereLabs/aya-expanse-8b
474
+ dtype: int64
475
+ - name: Qwen/Qwen3-8B
476
+ dtype: int64
477
+ - name: bigscience/bloom
478
+ dtype: int64
479
+ - name: common-pile/comma-v0.1-1t
480
+ dtype: int64
481
+ - name: facebook/xglm-564M
482
+ dtype: int64
483
+ - name: google-bert/bert-base-multilingual-cased
484
+ dtype: int64
485
+ - name: google/byt5-small
486
+ dtype: int64
487
+ - name: google/gemma-2-2b
488
+ dtype: int64
489
+ - name: gpt2
490
+ dtype: int64
491
+ - name: meta-llama/Llama-3.2-1B
492
+ dtype: int64
493
+ - name: microsoft/Phi-3-mini-4k-instruct
494
+ dtype: int64
495
+ - name: mistralai/tekken
496
+ dtype: int64
497
+ - name: tiktoken/gpt-4o
498
+ dtype: int64
499
+ - name: tokenmonster/englishcode-32000-consistent-v1
500
+ dtype: int64
501
+ splits:
502
+ - name: test
503
+ num_bytes: 566
504
+ num_examples: 1
505
+ download_size: 27881
506
+ dataset_size: 566
507
  configs:
508
  - config_name: tokenizer_robustness_completion_stem_canonical
509
  data_files:
 
517
  data_files:
518
  - split: test
519
  path: tokenizer_robustness_completion_stem_colloquial/test-*
520
+ - config_name: tokenizer_robustness_completion_stem_compounds
521
+ data_files:
522
+ - split: test
523
+ path: tokenizer_robustness_completion_stem_compounds/test-*
524
  ---
525
 
526
  # Dataset Card for Tokenization Robustness
tokenizer_robustness_completion_stem_compounds/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f432975522bf02d96d3a7f3355fd5c9471ceff98c4f6552ddab837373535c51b
3
+ size 27881