gsaltintas commited on
Commit
e0ab80c
·
verified ·
1 Parent(s): bb04787

Uploading tokenizer_robustness_completion_english_spelled_out subset

Browse files
README.md CHANGED
@@ -3108,6 +3108,130 @@ dataset_info:
3108
  num_examples: 40
3109
  download_size: 41512
3110
  dataset_size: 23738
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3111
  configs:
3112
  - config_name: tokenizer_robustness_completion_english_abbreviations
3113
  data_files:
@@ -3209,6 +3333,10 @@ configs:
3209
  data_files:
3210
  - split: test
3211
  path: tokenizer_robustness_completion_english_spaced_styling/test-*
 
 
 
 
3212
  ---
3213
 
3214
  # Dataset Card for Tokenization Robustness
 
3108
  num_examples: 40
3109
  download_size: 41512
3110
  dataset_size: 23738
3111
+ - config_name: tokenizer_robustness_completion_english_spelled_out
3112
+ features:
3113
+ - name: question
3114
+ dtype: string
3115
+ - name: choices
3116
+ list: string
3117
+ - name: answer
3118
+ dtype: int64
3119
+ - name: answer_label
3120
+ dtype: string
3121
+ - name: split
3122
+ dtype: string
3123
+ - name: subcategories
3124
+ dtype: string
3125
+ - name: category
3126
+ dtype: string
3127
+ - name: lang
3128
+ dtype: string
3129
+ - name: second_lang
3130
+ dtype: string
3131
+ - name: notes
3132
+ dtype: string
3133
+ - name: id
3134
+ dtype: string
3135
+ - name: set_id
3136
+ dtype: string
3137
+ - name: variation_id
3138
+ dtype: string
3139
+ - name: vanilla_cos_sim_to_canonical
3140
+ struct:
3141
+ - name: CohereLabs/aya-expanse-8b
3142
+ dtype: float64
3143
+ - name: Qwen/Qwen3-8B
3144
+ dtype: float64
3145
+ - name: bigscience/bloom
3146
+ dtype: float64
3147
+ - name: common-pile/comma-v0.1-1t
3148
+ dtype: float64
3149
+ - name: facebook/xglm-564M
3150
+ dtype: float64
3151
+ - name: google-bert/bert-base-multilingual-cased
3152
+ dtype: float64
3153
+ - name: google/byt5-small
3154
+ dtype: float64
3155
+ - name: google/gemma-2-2b
3156
+ dtype: float64
3157
+ - name: gpt2
3158
+ dtype: float64
3159
+ - name: meta-llama/Llama-3.2-1B
3160
+ dtype: float64
3161
+ - name: microsoft/Phi-3-mini-4k-instruct
3162
+ dtype: float64
3163
+ - name: mistralai/tekken
3164
+ dtype: float64
3165
+ - name: tiktoken/gpt-4o
3166
+ dtype: float64
3167
+ - name: tokenmonster/englishcode-32000-consistent-v1
3168
+ dtype: float64
3169
+ - name: trimmed_cos_sim_to_canonical
3170
+ struct:
3171
+ - name: CohereLabs/aya-expanse-8b
3172
+ dtype: float64
3173
+ - name: Qwen/Qwen3-8B
3174
+ dtype: float64
3175
+ - name: bigscience/bloom
3176
+ dtype: float64
3177
+ - name: common-pile/comma-v0.1-1t
3178
+ dtype: float64
3179
+ - name: facebook/xglm-564M
3180
+ dtype: float64
3181
+ - name: google-bert/bert-base-multilingual-cased
3182
+ dtype: float64
3183
+ - name: google/byt5-small
3184
+ dtype: float64
3185
+ - name: google/gemma-2-2b
3186
+ dtype: float64
3187
+ - name: gpt2
3188
+ dtype: float64
3189
+ - name: meta-llama/Llama-3.2-1B
3190
+ dtype: float64
3191
+ - name: microsoft/Phi-3-mini-4k-instruct
3192
+ dtype: float64
3193
+ - name: mistralai/tekken
3194
+ dtype: float64
3195
+ - name: tiktoken/gpt-4o
3196
+ dtype: float64
3197
+ - name: tokenmonster/englishcode-32000-consistent-v1
3198
+ dtype: float64
3199
+ - name: token_counts
3200
+ struct:
3201
+ - name: CohereLabs/aya-expanse-8b
3202
+ dtype: int64
3203
+ - name: Qwen/Qwen3-8B
3204
+ dtype: int64
3205
+ - name: bigscience/bloom
3206
+ dtype: int64
3207
+ - name: common-pile/comma-v0.1-1t
3208
+ dtype: int64
3209
+ - name: facebook/xglm-564M
3210
+ dtype: int64
3211
+ - name: google-bert/bert-base-multilingual-cased
3212
+ dtype: int64
3213
+ - name: google/byt5-small
3214
+ dtype: int64
3215
+ - name: google/gemma-2-2b
3216
+ dtype: int64
3217
+ - name: gpt2
3218
+ dtype: int64
3219
+ - name: meta-llama/Llama-3.2-1B
3220
+ dtype: int64
3221
+ - name: microsoft/Phi-3-mini-4k-instruct
3222
+ dtype: int64
3223
+ - name: mistralai/tekken
3224
+ dtype: int64
3225
+ - name: tiktoken/gpt-4o
3226
+ dtype: int64
3227
+ - name: tokenmonster/englishcode-32000-consistent-v1
3228
+ dtype: int64
3229
+ splits:
3230
+ - name: test
3231
+ num_bytes: 8091
3232
+ num_examples: 15
3233
+ download_size: 30689
3234
+ dataset_size: 8091
3235
  configs:
3236
  - config_name: tokenizer_robustness_completion_english_abbreviations
3237
  data_files:
 
3333
  data_files:
3334
  - split: test
3335
  path: tokenizer_robustness_completion_english_spaced_styling/test-*
3336
+ - config_name: tokenizer_robustness_completion_english_spelled_out
3337
+ data_files:
3338
+ - split: test
3339
+ path: tokenizer_robustness_completion_english_spelled_out/test-*
3340
  ---
3341
 
3342
  # Dataset Card for Tokenization Robustness
tokenizer_robustness_completion_english_spelled_out/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37ba481128c1d9c93ec00461e1ff46e5a3e4e80f3cba30b0b9ea986c779bdce7
3
+ size 30689