question
stringlengths 14
72
| choices
listlengths 4
4
| answer
int64 0
3
| answer_label
stringclasses 4
values | split
stringclasses 1
value | subcategories
stringclasses 2
values | category
stringclasses 2
values | lang
stringclasses 1
value | second_lang
stringclasses 1
value | notes
stringclasses 2
values | id
stringlengths 7
7
| set_id
stringlengths 5
5
| variation_id
stringclasses 7
values | vanilla_cos_sim_to_canonical
dict | trimmed_cos_sim_to_canonical
dict | token_counts
dict |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Dr Smith is a doctor. Occupation of Dr Smith is
|
[
"teach",
"judge",
"law",
"doc"
] | 3 |
D
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
300-0.2
|
300.0
|
0.2
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1.0000001192092896,
"common-pile/comma-v0.1-1t": 0.9999998211860657,
"facebook/xglm-564M": 1.0000001192092896,
"google-bert/bert-base-multilingual-cased": 0.9999998211860657,
"google/byt5-small": 1.0000001192092896,
"google/gemma-2-2b": 0.9999998807907104,
"gpt2": 0.9999998211860657,
"meta-llama/Llama-3.2-1B": 1.0000001192092896,
"microsoft/Phi-3-mini-4k-instruct": 0.9999998807907104,
"mistralai/tekken": 0.9999998807907104,
"tiktoken/gpt-4o": 0.9999999403953552,
"tokenmonster/englishcode-32000-consistent-v1": 1.0000001192092896
}
|
{
"CohereLabs/aya-expanse-8b": 1.0000003576278687,
"Qwen/Qwen3-8B": 1.0000004768371582,
"bigscience/bloom": 1.000000238418579,
"common-pile/comma-v0.1-1t": 1.0000004768371582,
"facebook/xglm-564M": 1.000000238418579,
"google-bert/bert-base-multilingual-cased": 1.0000004768371582,
"google/byt5-small": 1.0000004768371582,
"google/gemma-2-2b": 1.0000003576278687,
"gpt2": 1.0000004768371582,
"meta-llama/Llama-3.2-1B": 1.0000003576278687,
"microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687,
"mistralai/tekken": 1.0000004768371582,
"tiktoken/gpt-4o": 1.0000004768371582,
"tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687
}
|
{
"CohereLabs/aya-expanse-8b": 11,
"Qwen/Qwen3-8B": 11,
"bigscience/bloom": 12,
"common-pile/comma-v0.1-1t": 13,
"facebook/xglm-564M": 13,
"google-bert/bert-base-multilingual-cased": 14,
"google/byt5-small": 47,
"google/gemma-2-2b": 11,
"gpt2": 12,
"meta-llama/Llama-3.2-1B": 11,
"microsoft/Phi-3-mini-4k-instruct": 13,
"mistralai/tekken": 11,
"tiktoken/gpt-4o": 12,
"tokenmonster/englishcode-32000-consistent-v1": 11
}
|
||
Dr Smith is a doctor. Occupation of Dr Smith is
|
[
"Prof",
"MD",
"Hon",
"Esq"
] | 1 |
B
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
300-0.3
|
300.0
|
0.3
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 1.0000001192092896,
"common-pile/comma-v0.1-1t": 0.9999998211860657,
"facebook/xglm-564M": 1.0000001192092896,
"google-bert/bert-base-multilingual-cased": 0.9999998211860657,
"google/byt5-small": 1.0000001192092896,
"google/gemma-2-2b": 0.9999998807907104,
"gpt2": 0.9999998211860657,
"meta-llama/Llama-3.2-1B": 1.0000001192092896,
"microsoft/Phi-3-mini-4k-instruct": 0.9999998807907104,
"mistralai/tekken": 0.9999998807907104,
"tiktoken/gpt-4o": 0.9999999403953552,
"tokenmonster/englishcode-32000-consistent-v1": 1.0000001192092896
}
|
{
"CohereLabs/aya-expanse-8b": 1.0000003576278687,
"Qwen/Qwen3-8B": 1.0000004768371582,
"bigscience/bloom": 1.000000238418579,
"common-pile/comma-v0.1-1t": 1.0000004768371582,
"facebook/xglm-564M": 1.000000238418579,
"google-bert/bert-base-multilingual-cased": 1.0000004768371582,
"google/byt5-small": 1.0000004768371582,
"google/gemma-2-2b": 1.0000003576278687,
"gpt2": 1.0000004768371582,
"meta-llama/Llama-3.2-1B": 1.0000003576278687,
"microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687,
"mistralai/tekken": 1.0000004768371582,
"tiktoken/gpt-4o": 1.0000004768371582,
"tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687
}
|
{
"CohereLabs/aya-expanse-8b": 11,
"Qwen/Qwen3-8B": 11,
"bigscience/bloom": 12,
"common-pile/comma-v0.1-1t": 13,
"facebook/xglm-564M": 13,
"google-bert/bert-base-multilingual-cased": 14,
"google/byt5-small": 47,
"google/gemma-2-2b": 11,
"gpt2": 12,
"meta-llama/Llama-3.2-1B": 11,
"microsoft/Phi-3-mini-4k-instruct": 13,
"mistralai/tekken": 11,
"tiktoken/gpt-4o": 12,
"tokenmonster/englishcode-32000-consistent-v1": 11
}
|
||
Dr Smith is an MD. Occipation of Dr Smith is
|
[
"doctor",
"teacher",
"judge",
"lawyer"
] | 0 |
A
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
300-0.4
|
300.0
|
0.4
|
{
"CohereLabs/aya-expanse-8b": 0.8465678691864014,
"Qwen/Qwen3-8B": 0.8233134150505066,
"bigscience/bloom": 0.873887836933136,
"common-pile/comma-v0.1-1t": 0.7490841150283813,
"facebook/xglm-564M": 0.8408849239349365,
"google-bert/bert-base-multilingual-cased": 0.7985623478889465,
"google/byt5-small": 0.9447391033172607,
"google/gemma-2-2b": 0.835648775100708,
"gpt2": 0.8251903057098389,
"meta-llama/Llama-3.2-1B": 0.8439704775810242,
"microsoft/Phi-3-mini-4k-instruct": 0.8045406937599182,
"mistralai/tekken": 0.8300552368164062,
"tiktoken/gpt-4o": 0.8318107724189758,
"tokenmonster/englishcode-32000-consistent-v1": 0.7099841237068176
}
|
{
"CohereLabs/aya-expanse-8b": 0.37533700466156006,
"Qwen/Qwen3-8B": 0.23539593815803528,
"bigscience/bloom": 0.4065379500389099,
"common-pile/comma-v0.1-1t": 0.13768307864665985,
"facebook/xglm-564M": 0.3379814028739929,
"google-bert/bert-base-multilingual-cased": 0.4463389217853546,
"google/byt5-small": 0.6483171582221985,
"google/gemma-2-2b": 0.35768020153045654,
"gpt2": 0.3287959694862366,
"meta-llama/Llama-3.2-1B": 0.3190908432006836,
"microsoft/Phi-3-mini-4k-instruct": 0.2861732244491577,
"mistralai/tekken": 0.27699482440948486,
"tiktoken/gpt-4o": 0.3433341383934021,
"tokenmonster/englishcode-32000-consistent-v1": 0.19402597844600677
}
|
{
"CohereLabs/aya-expanse-8b": 12,
"Qwen/Qwen3-8B": 12,
"bigscience/bloom": 13,
"common-pile/comma-v0.1-1t": 15,
"facebook/xglm-564M": 13,
"google-bert/bert-base-multilingual-cased": 14,
"google/byt5-small": 44,
"google/gemma-2-2b": 12,
"gpt2": 12,
"meta-llama/Llama-3.2-1B": 12,
"microsoft/Phi-3-mini-4k-instruct": 14,
"mistralai/tekken": 12,
"tiktoken/gpt-4o": 12,
"tokenmonster/englishcode-32000-consistent-v1": 16
}
|
||
The price of this house is 1.03M dollars The cost of this house is
|
[
"1,030,000 dollars",
"1.030.000 dollars",
"1,030,000 dollars",
"1,030.000 dollars"
] | 0 |
A
|
test
|
Abbreviations, Unit combinations
|
Script / Orthography, Mathematical & Scientific Notation
|
eng_Latn
|
changed options to accommodate M
|
302-0.3
|
302.0
|
0.3
|
{
"CohereLabs/aya-expanse-8b": 0.9055666327476501,
"Qwen/Qwen3-8B": 0.9031217694282532,
"bigscience/bloom": 0.894109845161438,
"common-pile/comma-v0.1-1t": 0.9072467088699341,
"facebook/xglm-564M": 0.875411868095398,
"google-bert/bert-base-multilingual-cased": 0.8923179507255554,
"google/byt5-small": 0.9754881858825684,
"google/gemma-2-2b": 0.9033129215240479,
"gpt2": 0.8858630061149597,
"meta-llama/Llama-3.2-1B": 0.8717429041862488,
"microsoft/Phi-3-mini-4k-instruct": 0.8849417567253113,
"mistralai/tekken": 0.881121039390564,
"tiktoken/gpt-4o": 0.8818330764770508,
"tokenmonster/englishcode-32000-consistent-v1": 0.8387635350227356
}
|
{
"CohereLabs/aya-expanse-8b": 0.5830617547035217,
"Qwen/Qwen3-8B": 0.55271977186203,
"bigscience/bloom": 0.4354766607284546,
"common-pile/comma-v0.1-1t": 0.40262073278427124,
"facebook/xglm-564M": 0.28853747248649597,
"google-bert/bert-base-multilingual-cased": 0.42490333318710327,
"google/byt5-small": 0.6922744512557983,
"google/gemma-2-2b": 0.5299580097198486,
"gpt2": 0.3897722065448761,
"meta-llama/Llama-3.2-1B": 0.3892613649368286,
"microsoft/Phi-3-mini-4k-instruct": 0.5016849637031555,
"mistralai/tekken": 0.5062558054924011,
"tiktoken/gpt-4o": 0.37294676899909973,
"tokenmonster/englishcode-32000-consistent-v1": 0.31927287578582764
}
|
{
"CohereLabs/aya-expanse-8b": 19,
"Qwen/Qwen3-8B": 19,
"bigscience/bloom": 17,
"common-pile/comma-v0.1-1t": 20,
"facebook/xglm-564M": 16,
"google-bert/bert-base-multilingual-cased": 17,
"google/byt5-small": 66,
"google/gemma-2-2b": 19,
"gpt2": 17,
"meta-llama/Llama-3.2-1B": 18,
"microsoft/Phi-3-mini-4k-instruct": 19,
"mistralai/tekken": 19,
"tiktoken/gpt-4o": 18,
"tokenmonster/englishcode-32000-consistent-v1": 14
}
|
|
The # of continents on Earth is
|
[
"7",
"5",
"6",
"8"
] | 0 |
A
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
304-0.7
|
304.0
|
0.7
|
{
"CohereLabs/aya-expanse-8b": 0.8543572425842285,
"Qwen/Qwen3-8B": 0.8774767518043518,
"bigscience/bloom": 0.8670704960823059,
"common-pile/comma-v0.1-1t": 0.8822222352027893,
"facebook/xglm-564M": 0.8440374135971069,
"google-bert/bert-base-multilingual-cased": 0.8720088005065918,
"google/byt5-small": 0.9478123188018799,
"google/gemma-2-2b": 0.8665192127227783,
"gpt2": 0.8531822562217712,
"meta-llama/Llama-3.2-1B": 0.8680265545845032,
"microsoft/Phi-3-mini-4k-instruct": 0.8711134791374207,
"mistralai/tekken": 0.8590569496154785,
"tiktoken/gpt-4o": 0.8604445457458496,
"tokenmonster/englishcode-32000-consistent-v1": 0.7292405962944031
}
|
{
"CohereLabs/aya-expanse-8b": 0.15336325764656067,
"Qwen/Qwen3-8B": 0.17457246780395508,
"bigscience/bloom": 0.18650875985622406,
"common-pile/comma-v0.1-1t": 0.2236202359199524,
"facebook/xglm-564M": 0.05189888924360275,
"google-bert/bert-base-multilingual-cased": 0.11308808624744415,
"google/byt5-small": -0.02643284946680069,
"google/gemma-2-2b": 0.173978790640831,
"gpt2": 0.15644428133964539,
"meta-llama/Llama-3.2-1B": 0.19826795160770416,
"microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529,
"mistralai/tekken": 0.16212503612041473,
"tiktoken/gpt-4o": 0.1929726004600525,
"tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649
}
|
{
"CohereLabs/aya-expanse-8b": 7,
"Qwen/Qwen3-8B": 7,
"bigscience/bloom": 7,
"common-pile/comma-v0.1-1t": 8,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 31,
"google/gemma-2-2b": 7,
"gpt2": 7,
"meta-llama/Llama-3.2-1B": 7,
"microsoft/Phi-3-mini-4k-instruct": 8,
"mistralai/tekken": 7,
"tiktoken/gpt-4o": 7,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
||
The no. of continents on Earth is
|
[
"5",
"6",
"8",
"7"
] | 3 |
D
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
304-0.8
|
304.0
|
0.8
|
{
"CohereLabs/aya-expanse-8b": 0.7867254614830017,
"Qwen/Qwen3-8B": 0.824273943901062,
"bigscience/bloom": 0.811360776424408,
"common-pile/comma-v0.1-1t": 0.8064047694206238,
"facebook/xglm-564M": 0.8018571138381958,
"google-bert/bert-base-multilingual-cased": 0.8182386159896851,
"google/byt5-small": 0.9467005133628845,
"google/gemma-2-2b": 0.804913341999054,
"gpt2": 0.7843954563140869,
"meta-llama/Llama-3.2-1B": 0.7866487503051758,
"microsoft/Phi-3-mini-4k-instruct": 0.8095418810844421,
"mistralai/tekken": 0.7949608564376831,
"tiktoken/gpt-4o": 0.7873102426528931,
"tokenmonster/englishcode-32000-consistent-v1": 0.7079726457595825
}
|
{
"CohereLabs/aya-expanse-8b": 0.07414855808019638,
"Qwen/Qwen3-8B": 0.03043564409017563,
"bigscience/bloom": 0.07808922976255417,
"common-pile/comma-v0.1-1t": -0.02923409827053547,
"facebook/xglm-564M": 0.06573980301618576,
"google-bert/bert-base-multilingual-cased": -0.015136616304516792,
"google/byt5-small": -0.020509906113147736,
"google/gemma-2-2b": 0.0652402713894844,
"gpt2": 0.008664418943226337,
"meta-llama/Llama-3.2-1B": -0.008986718952655792,
"microsoft/Phi-3-mini-4k-instruct": -0.02389761433005333,
"mistralai/tekken": 0.06866628676652908,
"tiktoken/gpt-4o": 0.03231961280107498,
"tokenmonster/englishcode-32000-consistent-v1": 0.13605159521102905
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 8,
"common-pile/comma-v0.1-1t": 9,
"facebook/xglm-564M": 9,
"google-bert/bert-base-multilingual-cased": 9,
"google/byt5-small": 33,
"google/gemma-2-2b": 8,
"gpt2": 8,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 9,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
||
The capital city of IR is
|
[
"Mashhad",
"Baghdad",
"Tehran",
"Isfahan"
] | 2 |
C
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
305-0.6
|
305.0
|
0.6
|
{
"CohereLabs/aya-expanse-8b": 0.8646732568740845,
"Qwen/Qwen3-8B": 0.8861888647079468,
"bigscience/bloom": 0.8826625347137451,
"common-pile/comma-v0.1-1t": 0.8774277567863464,
"facebook/xglm-564M": 0.868166446685791,
"google-bert/bert-base-multilingual-cased": 0.8571589589118958,
"google/byt5-small": 0.9628745317459106,
"google/gemma-2-2b": 0.875096321105957,
"gpt2": 0.8794789910316467,
"meta-llama/Llama-3.2-1B": 0.8609308004379272,
"microsoft/Phi-3-mini-4k-instruct": 0.864482581615448,
"mistralai/tekken": 0.851871907711029,
"tiktoken/gpt-4o": 0.8562781810760498,
"tokenmonster/englishcode-32000-consistent-v1": 0.7831379771232605
}
|
{
"CohereLabs/aya-expanse-8b": 0.1503104269504547,
"Qwen/Qwen3-8B": 0.22655273973941803,
"bigscience/bloom": 0.24181559681892395,
"common-pile/comma-v0.1-1t": 0.22090619802474976,
"facebook/xglm-564M": 0.1887182891368866,
"google-bert/bert-base-multilingual-cased": 0.15784157812595367,
"google/byt5-small": 0.24735668301582336,
"google/gemma-2-2b": 0.18847647309303284,
"gpt2": 0.25452691316604614,
"meta-llama/Llama-3.2-1B": 0.16874276101589203,
"microsoft/Phi-3-mini-4k-instruct": 0.13977181911468506,
"mistralai/tekken": 0.13231658935546875,
"tiktoken/gpt-4o": 0.1192522943019867,
"tokenmonster/englishcode-32000-consistent-v1": 0.10740450024604797
}
|
{
"CohereLabs/aya-expanse-8b": 6,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 7,
"facebook/xglm-564M": 6,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 25,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 6,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 6
}
|
||
The # of days in a week is
|
[
"7",
"5",
"6",
"8"
] | 0 |
A
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
306-0.5
|
306.0
|
0.5
|
{
"CohereLabs/aya-expanse-8b": 0.8769029378890991,
"Qwen/Qwen3-8B": 0.8977583050727844,
"bigscience/bloom": 0.8890756368637085,
"common-pile/comma-v0.1-1t": 0.9004440307617188,
"facebook/xglm-564M": 0.8400768041610718,
"google-bert/bert-base-multilingual-cased": 0.8637771010398865,
"google/byt5-small": 0.9125043749809265,
"google/gemma-2-2b": 0.8880384564399719,
"gpt2": 0.8734878897666931,
"meta-llama/Llama-3.2-1B": 0.8814516067504883,
"microsoft/Phi-3-mini-4k-instruct": 0.8667116165161133,
"mistralai/tekken": 0.8863945007324219,
"tiktoken/gpt-4o": 0.8898459672927856,
"tokenmonster/englishcode-32000-consistent-v1": 0.758175253868103
}
|
{
"CohereLabs/aya-expanse-8b": 0.15336325764656067,
"Qwen/Qwen3-8B": 0.17457246780395508,
"bigscience/bloom": 0.18650875985622406,
"common-pile/comma-v0.1-1t": 0.2236202359199524,
"facebook/xglm-564M": 0.05189888924360275,
"google-bert/bert-base-multilingual-cased": 0.11308808624744415,
"google/byt5-small": -0.02643284946680069,
"google/gemma-2-2b": 0.173978790640831,
"gpt2": 0.15644428133964539,
"meta-llama/Llama-3.2-1B": 0.19826795160770416,
"microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529,
"mistralai/tekken": 0.16212503612041473,
"tiktoken/gpt-4o": 0.1929726004600525,
"tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 8,
"common-pile/comma-v0.1-1t": 10,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 26,
"google/gemma-2-2b": 8,
"gpt2": 8,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 8,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
||
The # of hours in a day is
|
[
"24",
"20",
"25",
"30"
] | 0 |
A
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
307-0.5
|
307.0
|
0.5
|
{
"CohereLabs/aya-expanse-8b": 0.8725788593292236,
"Qwen/Qwen3-8B": 0.8972941040992737,
"bigscience/bloom": 0.8836281299591064,
"common-pile/comma-v0.1-1t": 0.8991660475730896,
"facebook/xglm-564M": 0.8364160656929016,
"google-bert/bert-base-multilingual-cased": 0.862697422504425,
"google/byt5-small": 0.9182883501052856,
"google/gemma-2-2b": 0.8852136135101318,
"gpt2": 0.8709038496017456,
"meta-llama/Llama-3.2-1B": 0.8853892683982849,
"microsoft/Phi-3-mini-4k-instruct": 0.8643823862075806,
"mistralai/tekken": 0.8764235377311707,
"tiktoken/gpt-4o": 0.8882663249969482,
"tokenmonster/englishcode-32000-consistent-v1": 0.7297150492668152
}
|
{
"CohereLabs/aya-expanse-8b": 0.15336325764656067,
"Qwen/Qwen3-8B": 0.17457246780395508,
"bigscience/bloom": 0.18650875985622406,
"common-pile/comma-v0.1-1t": 0.2236202359199524,
"facebook/xglm-564M": 0.05189888924360275,
"google-bert/bert-base-multilingual-cased": 0.11308808624744415,
"google/byt5-small": -0.02643284946680069,
"google/gemma-2-2b": 0.173978790640831,
"gpt2": 0.15644428133964539,
"meta-llama/Llama-3.2-1B": 0.19826795160770416,
"microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529,
"mistralai/tekken": 0.16212503612041473,
"tiktoken/gpt-4o": 0.1929726004600525,
"tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 8,
"common-pile/comma-v0.1-1t": 10,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 26,
"google/gemma-2-2b": 8,
"gpt2": 8,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 8,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
||
The # of legs a cow has is
|
[
"8",
"4",
"3",
"5"
] | 1 |
B
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
308-0.5
|
308.0
|
0.5
|
{
"CohereLabs/aya-expanse-8b": 0.8784412145614624,
"Qwen/Qwen3-8B": 0.903361439704895,
"bigscience/bloom": 0.8900041580200195,
"common-pile/comma-v0.1-1t": 0.8913084268569946,
"facebook/xglm-564M": 0.8609603047370911,
"google-bert/bert-base-multilingual-cased": 0.8892489671707153,
"google/byt5-small": 0.9177855849266052,
"google/gemma-2-2b": 0.8813795447349548,
"gpt2": 0.877772331237793,
"meta-llama/Llama-3.2-1B": 0.8910291790962219,
"microsoft/Phi-3-mini-4k-instruct": 0.8654680848121643,
"mistralai/tekken": 0.8807762265205383,
"tiktoken/gpt-4o": 0.8883159160614014,
"tokenmonster/englishcode-32000-consistent-v1": 0.7710891962051392
}
|
{
"CohereLabs/aya-expanse-8b": 0.15336325764656067,
"Qwen/Qwen3-8B": 0.17457246780395508,
"bigscience/bloom": 0.18650875985622406,
"common-pile/comma-v0.1-1t": 0.2236202359199524,
"facebook/xglm-564M": 0.05189888924360275,
"google-bert/bert-base-multilingual-cased": 0.11308808624744415,
"google/byt5-small": -0.02643284946680069,
"google/gemma-2-2b": 0.173978790640831,
"gpt2": 0.15644428133964539,
"meta-llama/Llama-3.2-1B": 0.19826795160770416,
"microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529,
"mistralai/tekken": 0.16212503612041473,
"tiktoken/gpt-4o": 0.1929726004600525,
"tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 8,
"common-pile/comma-v0.1-1t": 9,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 9,
"google/byt5-small": 26,
"google/gemma-2-2b": 8,
"gpt2": 8,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 8,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 8
}
|
||
The # of minutes in 2 hours is
|
[
"100",
"140",
"120",
"90"
] | 2 |
C
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
309-0.5
|
309.0
|
0.5
|
{
"CohereLabs/aya-expanse-8b": 0.8936583995819092,
"Qwen/Qwen3-8B": 0.9028605222702026,
"bigscience/bloom": 0.901308000087738,
"common-pile/comma-v0.1-1t": 0.9184684753417969,
"facebook/xglm-564M": 0.8589028716087341,
"google-bert/bert-base-multilingual-cased": 0.8760905265808105,
"google/byt5-small": 0.9449029564857483,
"google/gemma-2-2b": 0.9074434041976929,
"gpt2": 0.8900853991508484,
"meta-llama/Llama-3.2-1B": 0.9088870286941528,
"microsoft/Phi-3-mini-4k-instruct": 0.8868902325630188,
"mistralai/tekken": 0.9065263271331787,
"tiktoken/gpt-4o": 0.9071903824806213,
"tokenmonster/englishcode-32000-consistent-v1": 0.7755128145217896
}
|
{
"CohereLabs/aya-expanse-8b": 0.15336325764656067,
"Qwen/Qwen3-8B": 0.17457246780395508,
"bigscience/bloom": 0.18650875985622406,
"common-pile/comma-v0.1-1t": 0.2236202359199524,
"facebook/xglm-564M": 0.05189888924360275,
"google-bert/bert-base-multilingual-cased": 0.11308808624744415,
"google/byt5-small": -0.02643284946680069,
"google/gemma-2-2b": 0.173978790640831,
"gpt2": 0.15644428133964539,
"meta-llama/Llama-3.2-1B": 0.19826795160770416,
"microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529,
"mistralai/tekken": 0.16212503612041473,
"tiktoken/gpt-4o": 0.1929726004600525,
"tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649
}
|
{
"CohereLabs/aya-expanse-8b": 9,
"Qwen/Qwen3-8B": 9,
"bigscience/bloom": 8,
"common-pile/comma-v0.1-1t": 11,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 30,
"google/gemma-2-2b": 9,
"gpt2": 8,
"meta-llama/Llama-3.2-1B": 9,
"microsoft/Phi-3-mini-4k-instruct": 9,
"mistralai/tekken": 9,
"tiktoken/gpt-4o": 9,
"tokenmonster/englishcode-32000-consistent-v1": 8
}
|
||
The # of months in a year is
|
[
"10",
"11",
"12",
"13"
] | 2 |
C
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
310-0.5
|
310.0
|
0.5
|
{
"CohereLabs/aya-expanse-8b": 0.8738687038421631,
"Qwen/Qwen3-8B": 0.8943588733673096,
"bigscience/bloom": 0.8882232308387756,
"common-pile/comma-v0.1-1t": 0.9044252634048462,
"facebook/xglm-564M": 0.8348524570465088,
"google-bert/bert-base-multilingual-cased": 0.8641276359558105,
"google/byt5-small": 0.9306106567382812,
"google/gemma-2-2b": 0.8897289633750916,
"gpt2": 0.8724748492240906,
"meta-llama/Llama-3.2-1B": 0.8823840022087097,
"microsoft/Phi-3-mini-4k-instruct": 0.8683964610099792,
"mistralai/tekken": 0.8803357481956482,
"tiktoken/gpt-4o": 0.8929274678230286,
"tokenmonster/englishcode-32000-consistent-v1": 0.7467564344406128
}
|
{
"CohereLabs/aya-expanse-8b": 0.15336325764656067,
"Qwen/Qwen3-8B": 0.17457246780395508,
"bigscience/bloom": 0.18650875985622406,
"common-pile/comma-v0.1-1t": 0.2236202359199524,
"facebook/xglm-564M": 0.05189888924360275,
"google-bert/bert-base-multilingual-cased": 0.11308808624744415,
"google/byt5-small": -0.02643284946680069,
"google/gemma-2-2b": 0.173978790640831,
"gpt2": 0.15644428133964539,
"meta-llama/Llama-3.2-1B": 0.19826795160770416,
"microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529,
"mistralai/tekken": 0.16212503612041473,
"tiktoken/gpt-4o": 0.1929726004600525,
"tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 8,
"common-pile/comma-v0.1-1t": 10,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 28,
"google/gemma-2-2b": 8,
"gpt2": 8,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 8,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
||
The # of seconds in a minute is
|
[
"60",
"50",
"100",
"30"
] | 0 |
A
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
311-0.5
|
311.0
|
0.5
|
{
"CohereLabs/aya-expanse-8b": 0.8620327711105347,
"Qwen/Qwen3-8B": 0.8879849910736084,
"bigscience/bloom": 0.8815665245056152,
"common-pile/comma-v0.1-1t": 0.8988855481147766,
"facebook/xglm-564M": 0.825326144695282,
"google-bert/bert-base-multilingual-cased": 0.8537793159484863,
"google/byt5-small": 0.9459080100059509,
"google/gemma-2-2b": 0.8843560218811035,
"gpt2": 0.8683614134788513,
"meta-llama/Llama-3.2-1B": 0.8789793848991394,
"microsoft/Phi-3-mini-4k-instruct": 0.8600611090660095,
"mistralai/tekken": 0.8747711777687073,
"tiktoken/gpt-4o": 0.8857784271240234,
"tokenmonster/englishcode-32000-consistent-v1": 0.7317348122596741
}
|
{
"CohereLabs/aya-expanse-8b": 0.15336325764656067,
"Qwen/Qwen3-8B": 0.17457246780395508,
"bigscience/bloom": 0.18650875985622406,
"common-pile/comma-v0.1-1t": 0.2236202359199524,
"facebook/xglm-564M": 0.05189888924360275,
"google-bert/bert-base-multilingual-cased": 0.11308808624744415,
"google/byt5-small": -0.02643284946680069,
"google/gemma-2-2b": 0.173978790640831,
"gpt2": 0.15644428133964539,
"meta-llama/Llama-3.2-1B": 0.19826795160770416,
"microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529,
"mistralai/tekken": 0.16212503612041473,
"tiktoken/gpt-4o": 0.1929726004600525,
"tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 8,
"common-pile/comma-v0.1-1t": 10,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 31,
"google/gemma-2-2b": 8,
"gpt2": 8,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 8,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
||
The # of sides a hexagon has is
|
[
"5",
"7",
"6",
"8"
] | 2 |
C
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
312-0.5
|
312.0
|
0.5
|
{
"CohereLabs/aya-expanse-8b": 0.862587571144104,
"Qwen/Qwen3-8B": 0.9136554002761841,
"bigscience/bloom": 0.9019626975059509,
"common-pile/comma-v0.1-1t": 0.897915244102478,
"facebook/xglm-564M": 0.904183030128479,
"google-bert/bert-base-multilingual-cased": 0.8995038270950317,
"google/byt5-small": 0.9488453269004822,
"google/gemma-2-2b": 0.8825085163116455,
"gpt2": 0.8939532041549683,
"meta-llama/Llama-3.2-1B": 0.9000940322875977,
"microsoft/Phi-3-mini-4k-instruct": 0.8842352628707886,
"mistralai/tekken": 0.8938607573509216,
"tiktoken/gpt-4o": 0.9065819382667542,
"tokenmonster/englishcode-32000-consistent-v1": 0.8501964807510376
}
|
{
"CohereLabs/aya-expanse-8b": 0.15336325764656067,
"Qwen/Qwen3-8B": 0.17457246780395508,
"bigscience/bloom": 0.18650875985622406,
"common-pile/comma-v0.1-1t": 0.2236202359199524,
"facebook/xglm-564M": 0.05189888924360275,
"google-bert/bert-base-multilingual-cased": 0.11308808624744415,
"google/byt5-small": -0.02643284946680069,
"google/gemma-2-2b": 0.173978790640831,
"gpt2": 0.15644428133964539,
"meta-llama/Llama-3.2-1B": 0.19826795160770416,
"microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529,
"mistralai/tekken": 0.16212503612041473,
"tiktoken/gpt-4o": 0.1929726004600525,
"tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 9,
"bigscience/bloom": 9,
"common-pile/comma-v0.1-1t": 10,
"facebook/xglm-564M": 10,
"google-bert/bert-base-multilingual-cased": 10,
"google/byt5-small": 31,
"google/gemma-2-2b": 8,
"gpt2": 9,
"meta-llama/Llama-3.2-1B": 9,
"microsoft/Phi-3-mini-4k-instruct": 9,
"mistralai/tekken": 9,
"tiktoken/gpt-4o": 9,
"tokenmonster/englishcode-32000-consistent-v1": 11
}
|
||
The # of sides a triangle has is
|
[
"2",
"4",
"3",
"5"
] | 2 |
C
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
313-0.4
|
313.0
|
0.4
|
{
"CohereLabs/aya-expanse-8b": 0.8708062171936035,
"Qwen/Qwen3-8B": 0.9027498364448547,
"bigscience/bloom": 0.8900589942932129,
"common-pile/comma-v0.1-1t": 0.8905171155929565,
"facebook/xglm-564M": 0.8620559573173523,
"google-bert/bert-base-multilingual-cased": 0.8632290959358215,
"google/byt5-small": 0.9473434090614319,
"google/gemma-2-2b": 0.8901822566986084,
"gpt2": 0.8811707496643066,
"meta-llama/Llama-3.2-1B": 0.8887614607810974,
"microsoft/Phi-3-mini-4k-instruct": 0.8653461933135986,
"mistralai/tekken": 0.8795601725578308,
"tiktoken/gpt-4o": 0.8902370929718018,
"tokenmonster/englishcode-32000-consistent-v1": 0.7794355750083923
}
|
{
"CohereLabs/aya-expanse-8b": 0.15336325764656067,
"Qwen/Qwen3-8B": 0.17457246780395508,
"bigscience/bloom": 0.18650875985622406,
"common-pile/comma-v0.1-1t": 0.2236202359199524,
"facebook/xglm-564M": 0.05189888924360275,
"google-bert/bert-base-multilingual-cased": 0.11308808624744415,
"google/byt5-small": -0.02643284946680069,
"google/gemma-2-2b": 0.173978790640831,
"gpt2": 0.15644428133964539,
"meta-llama/Llama-3.2-1B": 0.19826795160770416,
"microsoft/Phi-3-mini-4k-instruct": 0.0790470540523529,
"mistralai/tekken": 0.16212503612041473,
"tiktoken/gpt-4o": 0.1929726004600525,
"tokenmonster/englishcode-32000-consistent-v1": 0.1629045158624649
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 8,
"common-pile/comma-v0.1-1t": 9,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 32,
"google/gemma-2-2b": 8,
"gpt2": 8,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 8,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 8
}
|
||
In "I work at Apple", Apple is a
|
[
"pers.",
"cty.",
"fr.",
"co."
] | 3 |
D
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
314-0.3
|
314.0
|
0.3
|
{
"CohereLabs/aya-expanse-8b": 0.9999998807907104,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 0.9999998807907104,
"common-pile/comma-v0.1-1t": 0.9999998807907104,
"facebook/xglm-564M": 1.000000238418579,
"google-bert/bert-base-multilingual-cased": 1.0000001192092896,
"google/byt5-small": 1.0000001192092896,
"google/gemma-2-2b": 0.9999998211860657,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1.0000001192092896,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 0.9999998807907104,
"tokenmonster/englishcode-32000-consistent-v1": 0.9999997615814209
}
|
{
"CohereLabs/aya-expanse-8b": 0.9999999403953552,
"Qwen/Qwen3-8B": 1.000000238418579,
"bigscience/bloom": 1.0000003576278687,
"common-pile/comma-v0.1-1t": 1.0000004768371582,
"facebook/xglm-564M": 1.0000004768371582,
"google-bert/bert-base-multilingual-cased": 1.000000238418579,
"google/byt5-small": 1.000000238418579,
"google/gemma-2-2b": 1.000000238418579,
"gpt2": 1.0000003576278687,
"meta-llama/Llama-3.2-1B": 1.0000003576278687,
"microsoft/Phi-3-mini-4k-instruct": 1.0000004768371582,
"mistralai/tekken": 1.0000003576278687,
"tiktoken/gpt-4o": 1.0000003576278687,
"tokenmonster/englishcode-32000-consistent-v1": 1.0000005960464478
}
|
{
"CohereLabs/aya-expanse-8b": 10,
"Qwen/Qwen3-8B": 10,
"bigscience/bloom": 10,
"common-pile/comma-v0.1-1t": 10,
"facebook/xglm-564M": 10,
"google-bert/bert-base-multilingual-cased": 11,
"google/byt5-small": 32,
"google/gemma-2-2b": 10,
"gpt2": 10,
"meta-llama/Llama-3.2-1B": 10,
"microsoft/Phi-3-mini-4k-instruct": 10,
"mistralai/tekken": 10,
"tiktoken/gpt-4o": 10,
"tokenmonster/englishcode-32000-consistent-v1": 8
}
|
||
In "I work at Google", Google is a
|
[
"co.",
"pers.",
"cty.",
"fr."
] | 0 |
A
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
315-0.4
|
315.0
|
0.4
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 0.9999998807907104,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1.0000001192092896,
"google/byt5-small": 1.0000001192092896,
"google/gemma-2-2b": 0.9999998807907104,
"gpt2": 1.0000001192092896,
"meta-llama/Llama-3.2-1B": 0.9999999403953552,
"microsoft/Phi-3-mini-4k-instruct": 1.0000001192092896,
"mistralai/tekken": 0.9999998807907104,
"tiktoken/gpt-4o": 1.0000001192092896,
"tokenmonster/englishcode-32000-consistent-v1": 0.9999997615814209
}
|
{
"CohereLabs/aya-expanse-8b": 0.9999999403953552,
"Qwen/Qwen3-8B": 1.000000238418579,
"bigscience/bloom": 1.0000003576278687,
"common-pile/comma-v0.1-1t": 1.0000004768371582,
"facebook/xglm-564M": 1.0000004768371582,
"google-bert/bert-base-multilingual-cased": 1.000000238418579,
"google/byt5-small": 1.000000238418579,
"google/gemma-2-2b": 1.000000238418579,
"gpt2": 1.0000003576278687,
"meta-llama/Llama-3.2-1B": 1.0000003576278687,
"microsoft/Phi-3-mini-4k-instruct": 1.0000004768371582,
"mistralai/tekken": 1.0000003576278687,
"tiktoken/gpt-4o": 1.0000003576278687,
"tokenmonster/englishcode-32000-consistent-v1": 1.0000005960464478
}
|
{
"CohereLabs/aya-expanse-8b": 10,
"Qwen/Qwen3-8B": 10,
"bigscience/bloom": 10,
"common-pile/comma-v0.1-1t": 10,
"facebook/xglm-564M": 10,
"google-bert/bert-base-multilingual-cased": 11,
"google/byt5-small": 34,
"google/gemma-2-2b": 10,
"gpt2": 10,
"meta-llama/Llama-3.2-1B": 10,
"microsoft/Phi-3-mini-4k-instruct": 10,
"mistralai/tekken": 10,
"tiktoken/gpt-4o": 10,
"tokenmonster/englishcode-32000-consistent-v1": 8
}
|
||
In "Microsoft released a new update", Microsoft is a
|
[
"pers.",
"cty.",
"co.",
"fr."
] | 2 |
C
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
316-0.5
|
316.0
|
0.5
|
{
"CohereLabs/aya-expanse-8b": 0.9999996423721313,
"Qwen/Qwen3-8B": 0.9999999403953552,
"bigscience/bloom": 0.9999998807907104,
"common-pile/comma-v0.1-1t": 1.0000001192092896,
"facebook/xglm-564M": 0.9999998211860657,
"google-bert/bert-base-multilingual-cased": 0.9999999403953552,
"google/byt5-small": 1,
"google/gemma-2-2b": 0.9999996423721313,
"gpt2": 1.000000238418579,
"meta-llama/Llama-3.2-1B": 0.9999999403953552,
"microsoft/Phi-3-mini-4k-instruct": 1.0000001192092896,
"mistralai/tekken": 1.0000001192092896,
"tiktoken/gpt-4o": 1.0000001192092896,
"tokenmonster/englishcode-32000-consistent-v1": 1.0000001192092896
}
|
{
"CohereLabs/aya-expanse-8b": 0.9999999403953552,
"Qwen/Qwen3-8B": 1.000000238418579,
"bigscience/bloom": 1.0000003576278687,
"common-pile/comma-v0.1-1t": 1.0000004768371582,
"facebook/xglm-564M": 1.0000004768371582,
"google-bert/bert-base-multilingual-cased": 1.000000238418579,
"google/byt5-small": 1.000000238418579,
"google/gemma-2-2b": 1.000000238418579,
"gpt2": 1.0000003576278687,
"meta-llama/Llama-3.2-1B": 1.0000003576278687,
"microsoft/Phi-3-mini-4k-instruct": 1.0000004768371582,
"mistralai/tekken": 1.0000003576278687,
"tiktoken/gpt-4o": 1.0000003576278687,
"tokenmonster/englishcode-32000-consistent-v1": 1.0000005960464478
}
|
{
"CohereLabs/aya-expanse-8b": 11,
"Qwen/Qwen3-8B": 11,
"bigscience/bloom": 12,
"common-pile/comma-v0.1-1t": 11,
"facebook/xglm-564M": 11,
"google-bert/bert-base-multilingual-cased": 12,
"google/byt5-small": 52,
"google/gemma-2-2b": 11,
"gpt2": 11,
"meta-llama/Llama-3.2-1B": 11,
"microsoft/Phi-3-mini-4k-instruct": 11,
"mistralai/tekken": 11,
"tiktoken/gpt-4o": 11,
"tokenmonster/englishcode-32000-consistent-v1": 9
}
|
||
In "The cat sat on the mat", the subj. is
|
[
"the cat",
"sat",
"the mat",
"on"
] | 0 |
A
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
317-0.5
|
317.0
|
0.5
|
{
"CohereLabs/aya-expanse-8b": 0.9349287152290344,
"Qwen/Qwen3-8B": 0.9383683204650879,
"bigscience/bloom": 0.9273239374160767,
"common-pile/comma-v0.1-1t": 0.8940730094909668,
"facebook/xglm-564M": 0.8844199776649475,
"google-bert/bert-base-multilingual-cased": 0.8985786437988281,
"google/byt5-small": 0.9816980957984924,
"google/gemma-2-2b": 0.9320594668388367,
"gpt2": 0.8804003000259399,
"meta-llama/Llama-3.2-1B": 0.92516028881073,
"microsoft/Phi-3-mini-4k-instruct": 0.8588052988052368,
"mistralai/tekken": 0.8823502659797668,
"tiktoken/gpt-4o": 0.9314396381378174,
"tokenmonster/englishcode-32000-consistent-v1": 0.8803983330726624
}
|
{
"CohereLabs/aya-expanse-8b": 0.0772608071565628,
"Qwen/Qwen3-8B": 0.10374213010072708,
"bigscience/bloom": 0.08070938289165497,
"common-pile/comma-v0.1-1t": 0.07346736639738083,
"facebook/xglm-564M": 0.05824323743581772,
"google-bert/bert-base-multilingual-cased": 0.08166754245758057,
"google/byt5-small": -0.046926841139793396,
"google/gemma-2-2b": 0.0717279389500618,
"gpt2": 0.026471275836229324,
"meta-llama/Llama-3.2-1B": 0.06531772017478943,
"microsoft/Phi-3-mini-4k-instruct": -0.03549063205718994,
"mistralai/tekken": 0.07283864170312881,
"tiktoken/gpt-4o": 0.07994227856397629,
"tokenmonster/englishcode-32000-consistent-v1": 0.05385168641805649
}
|
{
"CohereLabs/aya-expanse-8b": 13,
"Qwen/Qwen3-8B": 13,
"bigscience/bloom": 13,
"common-pile/comma-v0.1-1t": 14,
"facebook/xglm-564M": 14,
"google-bert/bert-base-multilingual-cased": 15,
"google/byt5-small": 41,
"google/gemma-2-2b": 13,
"gpt2": 14,
"meta-llama/Llama-3.2-1B": 13,
"microsoft/Phi-3-mini-4k-instruct": 14,
"mistralai/tekken": 14,
"tiktoken/gpt-4o": 13,
"tokenmonster/englishcode-32000-consistent-v1": 12
}
|
||
The gas humans need to breathe to live is
|
[
"CH₄",
"O₂",
"He",
"H₂"
] | 1 |
B
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
322-0.5
|
322.0
|
0.5
|
{
"CohereLabs/aya-expanse-8b": 1.000000238418579,
"Qwen/Qwen3-8B": 1.0000001192092896,
"bigscience/bloom": 1.0000001192092896,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 1.0000001192092896,
"google-bert/bert-base-multilingual-cased": 0.9999999403953552,
"google/byt5-small": 0.9999997615814209,
"google/gemma-2-2b": 1.0000001192092896,
"gpt2": 1.0000001192092896,
"meta-llama/Llama-3.2-1B": 1.0000001192092896,
"microsoft/Phi-3-mini-4k-instruct": 1.0000001192092896,
"mistralai/tekken": 1.000000238418579,
"tiktoken/gpt-4o": 1.000000238418579,
"tokenmonster/englishcode-32000-consistent-v1": 0.9999999403953552
}
|
{
"CohereLabs/aya-expanse-8b": 1.0000003576278687,
"Qwen/Qwen3-8B": 1.0000004768371582,
"bigscience/bloom": 1.000000238418579,
"common-pile/comma-v0.1-1t": 1.0000004768371582,
"facebook/xglm-564M": 1.000000238418579,
"google-bert/bert-base-multilingual-cased": 1.0000004768371582,
"google/byt5-small": 1.0000004768371582,
"google/gemma-2-2b": 1.0000003576278687,
"gpt2": 1.0000004768371582,
"meta-llama/Llama-3.2-1B": 1.0000003576278687,
"microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687,
"mistralai/tekken": 1.0000004768371582,
"tiktoken/gpt-4o": 1.0000004768371582,
"tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687
}
|
{
"CohereLabs/aya-expanse-8b": 9,
"Qwen/Qwen3-8B": 9,
"bigscience/bloom": 9,
"common-pile/comma-v0.1-1t": 14,
"facebook/xglm-564M": 9,
"google-bert/bert-base-multilingual-cased": 10,
"google/byt5-small": 41,
"google/gemma-2-2b": 9,
"gpt2": 9,
"meta-llama/Llama-3.2-1B": 9,
"microsoft/Phi-3-mini-4k-instruct": 11,
"mistralai/tekken": 9,
"tiktoken/gpt-4o": 9,
"tokenmonster/englishcode-32000-consistent-v1": 8
}
|
||
Chad's cap. is
|
[
"Moundou",
"Abéché",
"Ngama",
"N'Djamena"
] | 3 |
D
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
326-0.3
|
326.0
|
0.3
|
{
"CohereLabs/aya-expanse-8b": 0.696435809135437,
"Qwen/Qwen3-8B": 0.7792327404022217,
"bigscience/bloom": 0.7729296684265137,
"common-pile/comma-v0.1-1t": 0.7977029085159302,
"facebook/xglm-564M": 0.7329659461975098,
"google-bert/bert-base-multilingual-cased": 0.7466813325881958,
"google/byt5-small": 0.9061909317970276,
"google/gemma-2-2b": 0.7286580204963684,
"gpt2": 0.783542275428772,
"meta-llama/Llama-3.2-1B": 0.7782202959060669,
"microsoft/Phi-3-mini-4k-instruct": 0.8116836547851562,
"mistralai/tekken": 0.7620350122451782,
"tiktoken/gpt-4o": 0.781197190284729,
"tokenmonster/englishcode-32000-consistent-v1": 0.8143059015274048
}
|
{
"CohereLabs/aya-expanse-8b": 0.1356840431690216,
"Qwen/Qwen3-8B": 0.054030366241931915,
"bigscience/bloom": 0.02676394209265709,
"common-pile/comma-v0.1-1t": 0.05686108395457268,
"facebook/xglm-564M": 0.06816975772380829,
"google-bert/bert-base-multilingual-cased": 0.057269658893346786,
"google/byt5-small": -0.02295071817934513,
"google/gemma-2-2b": 0.09922298789024353,
"gpt2": 0.10866591334342957,
"meta-llama/Llama-3.2-1B": 0.09607133269309998,
"microsoft/Phi-3-mini-4k-instruct": 0.13172483444213867,
"mistralai/tekken": 0.04374687373638153,
"tiktoken/gpt-4o": 0.08728465437889099,
"tokenmonster/englishcode-32000-consistent-v1": 0.06787453591823578
}
|
{
"CohereLabs/aya-expanse-8b": 5,
"Qwen/Qwen3-8B": 6,
"bigscience/bloom": 6,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 6,
"google-bert/bert-base-multilingual-cased": 6,
"google/byt5-small": 14,
"google/gemma-2-2b": 6,
"gpt2": 6,
"meta-llama/Llama-3.2-1B": 6,
"microsoft/Phi-3-mini-4k-instruct": 7,
"mistralai/tekken": 6,
"tiktoken/gpt-4o": 6,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
||
The capital of FR is
|
[
"London",
"Berlin",
"Rome",
"Paris"
] | 3 |
D
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
327-0.4
|
327.0
|
0.4
|
{
"CohereLabs/aya-expanse-8b": 0.7995725274085999,
"Qwen/Qwen3-8B": 0.8451246023178101,
"bigscience/bloom": 0.8283455967903137,
"common-pile/comma-v0.1-1t": 0.8237259984016418,
"facebook/xglm-564M": 0.806121826171875,
"google-bert/bert-base-multilingual-cased": 0.7981747388839722,
"google/byt5-small": 0.9172342419624329,
"google/gemma-2-2b": 0.7995125651359558,
"gpt2": 0.8134028315544128,
"meta-llama/Llama-3.2-1B": 0.8076421618461609,
"microsoft/Phi-3-mini-4k-instruct": 0.8156378865242004,
"mistralai/tekken": 0.7921538352966309,
"tiktoken/gpt-4o": 0.7901878356933594,
"tokenmonster/englishcode-32000-consistent-v1": 0.7765400409698486
}
|
{
"CohereLabs/aya-expanse-8b": 0.08211846649646759,
"Qwen/Qwen3-8B": 0.14105814695358276,
"bigscience/bloom": 0.1934700310230255,
"common-pile/comma-v0.1-1t": 0.09907892346382141,
"facebook/xglm-564M": 0.08311900496482849,
"google-bert/bert-base-multilingual-cased": 0.13695181906223297,
"google/byt5-small": 0.14798158407211304,
"google/gemma-2-2b": 0.01552680041640997,
"gpt2": 0.14244207739830017,
"meta-llama/Llama-3.2-1B": 0.14876556396484375,
"microsoft/Phi-3-mini-4k-instruct": 0.1234123557806015,
"mistralai/tekken": 0.08589836210012436,
"tiktoken/gpt-4o": 0.07952668517827988,
"tokenmonster/englishcode-32000-consistent-v1": 0.026012370362877846
}
|
{
"CohereLabs/aya-expanse-8b": 5,
"Qwen/Qwen3-8B": 5,
"bigscience/bloom": 5,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 5,
"google-bert/bert-base-multilingual-cased": 5,
"google/byt5-small": 20,
"google/gemma-2-2b": 5,
"gpt2": 5,
"meta-llama/Llama-3.2-1B": 5,
"microsoft/Phi-3-mini-4k-instruct": 5,
"mistralai/tekken": 5,
"tiktoken/gpt-4o": 5,
"tokenmonster/englishcode-32000-consistent-v1": 6
}
|
||
The capital of JP is
|
[
"Kyoto",
"Tokyo",
"Osaka",
"Hiroshima"
] | 1 |
B
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
328-0.4
|
328.0
|
0.4
|
{
"CohereLabs/aya-expanse-8b": 0.831386148929596,
"Qwen/Qwen3-8B": 0.8522142171859741,
"bigscience/bloom": 0.8232855200767517,
"common-pile/comma-v0.1-1t": 0.8368542194366455,
"facebook/xglm-564M": 0.8285537362098694,
"google-bert/bert-base-multilingual-cased": 0.8010603785514832,
"google/byt5-small": 0.9203615784645081,
"google/gemma-2-2b": 0.8251798152923584,
"gpt2": 0.8022357225418091,
"meta-llama/Llama-3.2-1B": 0.8143433332443237,
"microsoft/Phi-3-mini-4k-instruct": 0.7175941467285156,
"mistralai/tekken": 0.8012662529945374,
"tiktoken/gpt-4o": 0.8201245069503784,
"tokenmonster/englishcode-32000-consistent-v1": 0.8073024749755859
}
|
{
"CohereLabs/aya-expanse-8b": 0.1724080741405487,
"Qwen/Qwen3-8B": 0.12307695299386978,
"bigscience/bloom": 0.08853715658187866,
"common-pile/comma-v0.1-1t": 0.11716070026159286,
"facebook/xglm-564M": 0.15991993248462677,
"google-bert/bert-base-multilingual-cased": 0.08823803812265396,
"google/byt5-small": 0.2070769965648651,
"google/gemma-2-2b": 0.06976078450679779,
"gpt2": 0.08312174677848816,
"meta-llama/Llama-3.2-1B": 0.12662671506404877,
"microsoft/Phi-3-mini-4k-instruct": 0.13243405520915985,
"mistralai/tekken": 0.06049402803182602,
"tiktoken/gpt-4o": 0.14071625471115112,
"tokenmonster/englishcode-32000-consistent-v1": 0.10549987852573395
}
|
{
"CohereLabs/aya-expanse-8b": 5,
"Qwen/Qwen3-8B": 5,
"bigscience/bloom": 5,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 5,
"google-bert/bert-base-multilingual-cased": 5,
"google/byt5-small": 20,
"google/gemma-2-2b": 5,
"gpt2": 5,
"meta-llama/Llama-3.2-1B": 5,
"microsoft/Phi-3-mini-4k-instruct": 6,
"mistralai/tekken": 5,
"tiktoken/gpt-4o": 5,
"tokenmonster/englishcode-32000-consistent-v1": 6
}
|
||
The capital of TR is
|
[
"İstanbul",
"İzmir",
"Ankara",
"Bursa"
] | 2 |
C
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
329-0.3
|
329.0
|
0.3
|
{
"CohereLabs/aya-expanse-8b": 0.8062456846237183,
"Qwen/Qwen3-8B": 0.8371235132217407,
"bigscience/bloom": 0.8147814869880676,
"common-pile/comma-v0.1-1t": 0.8117464184761047,
"facebook/xglm-564M": 0.8058038949966431,
"google-bert/bert-base-multilingual-cased": 0.807705283164978,
"google/byt5-small": 0.8975408673286438,
"google/gemma-2-2b": 0.8226367831230164,
"gpt2": 0.7932828664779663,
"meta-llama/Llama-3.2-1B": 0.8048302531242371,
"microsoft/Phi-3-mini-4k-instruct": 0.7885987758636475,
"mistralai/tekken": 0.8118667602539062,
"tiktoken/gpt-4o": 0.7969956994056702,
"tokenmonster/englishcode-32000-consistent-v1": 0.7851678729057312
}
|
{
"CohereLabs/aya-expanse-8b": 0.08540067076683044,
"Qwen/Qwen3-8B": 0.09242145717144012,
"bigscience/bloom": 0.11954235285520554,
"common-pile/comma-v0.1-1t": 0.0688096210360527,
"facebook/xglm-564M": 0.056273337453603745,
"google-bert/bert-base-multilingual-cased": 0.184139221906662,
"google/byt5-small": 0.12901875376701355,
"google/gemma-2-2b": 0.11601769179105759,
"gpt2": 0.07733984291553497,
"meta-llama/Llama-3.2-1B": 0.14307504892349243,
"microsoft/Phi-3-mini-4k-instruct": -0.00042920373380184174,
"mistralai/tekken": 0.1253935694694519,
"tiktoken/gpt-4o": 0.04226153716444969,
"tokenmonster/englishcode-32000-consistent-v1": -0.0061051626689732075
}
|
{
"CohereLabs/aya-expanse-8b": 5,
"Qwen/Qwen3-8B": 5,
"bigscience/bloom": 5,
"common-pile/comma-v0.1-1t": 6,
"facebook/xglm-564M": 5,
"google-bert/bert-base-multilingual-cased": 5,
"google/byt5-small": 20,
"google/gemma-2-2b": 5,
"gpt2": 5,
"meta-llama/Llama-3.2-1B": 5,
"microsoft/Phi-3-mini-4k-instruct": 5,
"mistralai/tekken": 5,
"tiktoken/gpt-4o": 5,
"tokenmonster/englishcode-32000-consistent-v1": 6
}
|
||
The chem. formula for water is
|
[
"H2O",
"CO2",
"NaCl",
"O2"
] | 0 |
A
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
330-0.4
|
330.0
|
0.4
|
{
"CohereLabs/aya-expanse-8b": 0.8186859488487244,
"Qwen/Qwen3-8B": 0.8713029026985168,
"bigscience/bloom": 0.8499932885169983,
"common-pile/comma-v0.1-1t": 0.8503109216690063,
"facebook/xglm-564M": 0.7378727793693542,
"google-bert/bert-base-multilingual-cased": 0.7587925791740417,
"google/byt5-small": 0.9591166973114014,
"google/gemma-2-2b": 0.8184010982513428,
"gpt2": 0.8338674306869507,
"meta-llama/Llama-3.2-1B": 0.819198727607727,
"microsoft/Phi-3-mini-4k-instruct": 0.8195061683654785,
"mistralai/tekken": 0.8313064575195312,
"tiktoken/gpt-4o": 0.836508572101593,
"tokenmonster/englishcode-32000-consistent-v1": 0.848869264125824
}
|
{
"CohereLabs/aya-expanse-8b": 0.19135701656341553,
"Qwen/Qwen3-8B": 0.21313846111297607,
"bigscience/bloom": 0.23347987234592438,
"common-pile/comma-v0.1-1t": 0.18952646851539612,
"facebook/xglm-564M": 0.021598808467388153,
"google-bert/bert-base-multilingual-cased": 0.016012798994779587,
"google/byt5-small": -0.02367701567709446,
"google/gemma-2-2b": 0.17226368188858032,
"gpt2": 0.20595847070217133,
"meta-llama/Llama-3.2-1B": 0.2124955654144287,
"microsoft/Phi-3-mini-4k-instruct": 0.15150737762451172,
"mistralai/tekken": 0.24085702002048492,
"tiktoken/gpt-4o": 0.16686344146728516,
"tokenmonster/englishcode-32000-consistent-v1": 0.14736664295196533
}
|
{
"CohereLabs/aya-expanse-8b": 7,
"Qwen/Qwen3-8B": 7,
"bigscience/bloom": 7,
"common-pile/comma-v0.1-1t": 7,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 30,
"google/gemma-2-2b": 7,
"gpt2": 7,
"meta-llama/Llama-3.2-1B": 7,
"microsoft/Phi-3-mini-4k-instruct": 7,
"mistralai/tekken": 7,
"tiktoken/gpt-4o": 7,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
||
The intent in "What time does the store close?" is
|
[
"purch",
"book",
"complain",
"info"
] | 3 |
D
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
331-0.4
|
331.0
|
0.4
|
{
"CohereLabs/aya-expanse-8b": 0.9999998807907104,
"Qwen/Qwen3-8B": 1.0000001192092896,
"bigscience/bloom": 0.9999998807907104,
"common-pile/comma-v0.1-1t": 1.0000001192092896,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 1.0000001192092896,
"google/byt5-small": 1,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 0.9999995827674866,
"microsoft/Phi-3-mini-4k-instruct": 0.9999997615814209,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 0.9999998807907104,
"tokenmonster/englishcode-32000-consistent-v1": 0.9999999403953552
}
|
{
"CohereLabs/aya-expanse-8b": 1.0000003576278687,
"Qwen/Qwen3-8B": 1.0000004768371582,
"bigscience/bloom": 1.000000238418579,
"common-pile/comma-v0.1-1t": 1.0000004768371582,
"facebook/xglm-564M": 1.000000238418579,
"google-bert/bert-base-multilingual-cased": 1.0000004768371582,
"google/byt5-small": 1.0000004768371582,
"google/gemma-2-2b": 1.0000003576278687,
"gpt2": 1.0000004768371582,
"meta-llama/Llama-3.2-1B": 1.0000003576278687,
"microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687,
"mistralai/tekken": 1.0000004768371582,
"tiktoken/gpt-4o": 1.0000004768371582,
"tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687
}
|
{
"CohereLabs/aya-expanse-8b": 12,
"Qwen/Qwen3-8B": 12,
"bigscience/bloom": 12,
"common-pile/comma-v0.1-1t": 13,
"facebook/xglm-564M": 12,
"google-bert/bert-base-multilingual-cased": 13,
"google/byt5-small": 50,
"google/gemma-2-2b": 12,
"gpt2": 12,
"meta-llama/Llama-3.2-1B": 12,
"microsoft/Phi-3-mini-4k-instruct": 12,
"mistralai/tekken": 12,
"tiktoken/gpt-4o": 12,
"tokenmonster/englishcode-32000-consistent-v1": 11
}
|
||
The largest mammal in the world is
|
[
"dolphin",
"giraffe",
"blue whale",
"bear"
] | 2 |
C
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
332-0.5
|
332.0
|
0.5
|
{
"CohereLabs/aya-expanse-8b": 0.9999997615814209,
"Qwen/Qwen3-8B": 0.9999997615814209,
"bigscience/bloom": 0.9999997019767761,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 0.9999997615814209,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 0.9999998211860657,
"google/gemma-2-2b": 0.9999999403953552,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 0.9999998211860657,
"microsoft/Phi-3-mini-4k-instruct": 0.9999998807907104,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 0.9999997615814209,
"tokenmonster/englishcode-32000-consistent-v1": 0.9999998807907104
}
|
{
"CohereLabs/aya-expanse-8b": 1.0000003576278687,
"Qwen/Qwen3-8B": 1.0000004768371582,
"bigscience/bloom": 1.000000238418579,
"common-pile/comma-v0.1-1t": 1.0000004768371582,
"facebook/xglm-564M": 1.000000238418579,
"google-bert/bert-base-multilingual-cased": 1.0000004768371582,
"google/byt5-small": 1.0000004768371582,
"google/gemma-2-2b": 1.0000003576278687,
"gpt2": 1.0000004768371582,
"meta-llama/Llama-3.2-1B": 1.0000003576278687,
"microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687,
"mistralai/tekken": 1.0000004768371582,
"tiktoken/gpt-4o": 1.0000004768371582,
"tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687
}
|
{
"CohereLabs/aya-expanse-8b": 7,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 8,
"common-pile/comma-v0.1-1t": 8,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 34,
"google/gemma-2-2b": 7,
"gpt2": 7,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 9,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
||
The unit of measurement for temperature in the International System is
|
[
"°C",
"m",
"K",
"°R"
] | 2 |
C
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
333-0.4
|
333.0
|
0.4
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 1.0000001192092896,
"bigscience/bloom": 0.9999998807907104,
"common-pile/comma-v0.1-1t": 1,
"facebook/xglm-564M": 0.9999998807907104,
"google-bert/bert-base-multilingual-cased": 0.9999998211860657,
"google/byt5-small": 1,
"google/gemma-2-2b": 0.9999999403953552,
"gpt2": 1.0000001192092896,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1.0000001192092896,
"tiktoken/gpt-4o": 1.0000001192092896,
"tokenmonster/englishcode-32000-consistent-v1": 0.9999999403953552
}
|
{
"CohereLabs/aya-expanse-8b": 1.0000003576278687,
"Qwen/Qwen3-8B": 1.0000004768371582,
"bigscience/bloom": 1.000000238418579,
"common-pile/comma-v0.1-1t": 1.0000004768371582,
"facebook/xglm-564M": 1.000000238418579,
"google-bert/bert-base-multilingual-cased": 1.0000004768371582,
"google/byt5-small": 1.0000004768371582,
"google/gemma-2-2b": 1.0000003576278687,
"gpt2": 1.0000004768371582,
"meta-llama/Llama-3.2-1B": 1.0000003576278687,
"microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687,
"mistralai/tekken": 1.0000004768371582,
"tiktoken/gpt-4o": 1.0000004768371582,
"tokenmonster/englishcode-32000-consistent-v1": 1.000000238418579
}
|
{
"CohereLabs/aya-expanse-8b": 11,
"Qwen/Qwen3-8B": 11,
"bigscience/bloom": 11,
"common-pile/comma-v0.1-1t": 13,
"facebook/xglm-564M": 11,
"google-bert/bert-base-multilingual-cased": 11,
"google/byt5-small": 70,
"google/gemma-2-2b": 11,
"gpt2": 11,
"meta-llama/Llama-3.2-1B": 11,
"microsoft/Phi-3-mini-4k-instruct": 11,
"mistralai/tekken": 11,
"tiktoken/gpt-4o": 11,
"tokenmonster/englishcode-32000-consistent-v1": 10
}
|
||
The country whose space agency is NASA is
|
[
"RU",
"US",
"CN",
"JP"
] | 1 |
B
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
334-0.3
|
334.0
|
0.3
|
{
"CohereLabs/aya-expanse-8b": 0.9999997615814209,
"Qwen/Qwen3-8B": 0.9999998807907104,
"bigscience/bloom": 0.9999997019767761,
"common-pile/comma-v0.1-1t": 0.9999997615814209,
"facebook/xglm-564M": 0.9999997615814209,
"google-bert/bert-base-multilingual-cased": 1,
"google/byt5-small": 0.9999998211860657,
"google/gemma-2-2b": 0.9999997615814209,
"gpt2": 1.0000001192092896,
"meta-llama/Llama-3.2-1B": 0.9999998807907104,
"microsoft/Phi-3-mini-4k-instruct": 1.000000238418579,
"mistralai/tekken": 0.9999996423721313,
"tiktoken/gpt-4o": 0.9999998807907104,
"tokenmonster/englishcode-32000-consistent-v1": 1.0000001192092896
}
|
{
"CohereLabs/aya-expanse-8b": 1.0000003576278687,
"Qwen/Qwen3-8B": 1.0000004768371582,
"bigscience/bloom": 1.000000238418579,
"common-pile/comma-v0.1-1t": 1.0000004768371582,
"facebook/xglm-564M": 1.000000238418579,
"google-bert/bert-base-multilingual-cased": 1.0000004768371582,
"google/byt5-small": 1.0000004768371582,
"google/gemma-2-2b": 1.0000003576278687,
"gpt2": 1.0000004768371582,
"meta-llama/Llama-3.2-1B": 1.0000003576278687,
"microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687,
"mistralai/tekken": 1.0000004768371582,
"tiktoken/gpt-4o": 1.0000004768371582,
"tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 8,
"common-pile/comma-v0.1-1t": 8,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 41,
"google/gemma-2-2b": 8,
"gpt2": 8,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 9,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 9
}
|
||
The lang. spoken in Brazil is
|
[
"Spanish",
"French",
"Italian",
"Portuguese"
] | 3 |
D
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
335-0.5
|
335.0
|
0.5
|
{
"CohereLabs/aya-expanse-8b": 0.7913755178451538,
"Qwen/Qwen3-8B": 0.8113056421279907,
"bigscience/bloom": 0.8049988746643066,
"common-pile/comma-v0.1-1t": 0.8325293064117432,
"facebook/xglm-564M": 0.7717224359512329,
"google-bert/bert-base-multilingual-cased": 0.7581318616867065,
"google/byt5-small": 0.9475886821746826,
"google/gemma-2-2b": 0.7959599494934082,
"gpt2": 0.788682222366333,
"meta-llama/Llama-3.2-1B": 0.8063963055610657,
"microsoft/Phi-3-mini-4k-instruct": 0.7926681041717529,
"mistralai/tekken": 0.7794783711433411,
"tiktoken/gpt-4o": 0.8209499716758728,
"tokenmonster/englishcode-32000-consistent-v1": 0.7077274322509766
}
|
{
"CohereLabs/aya-expanse-8b": 0.10002417862415314,
"Qwen/Qwen3-8B": 0.07510721683502197,
"bigscience/bloom": 0.12772142887115479,
"common-pile/comma-v0.1-1t": 0.06116257607936859,
"facebook/xglm-564M": 0.09400343149900436,
"google-bert/bert-base-multilingual-cased": 0.024091584607958794,
"google/byt5-small": -0.025367457419633865,
"google/gemma-2-2b": 0.14458352327346802,
"gpt2": 0.10195167362689972,
"meta-llama/Llama-3.2-1B": 0.11951098591089249,
"microsoft/Phi-3-mini-4k-instruct": 0.04442334175109863,
"mistralai/tekken": 0.07316195219755173,
"tiktoken/gpt-4o": 0.16807381808757782,
"tokenmonster/englishcode-32000-consistent-v1": 0.16010284423828125
}
|
{
"CohereLabs/aya-expanse-8b": 7,
"Qwen/Qwen3-8B": 7,
"bigscience/bloom": 7,
"common-pile/comma-v0.1-1t": 8,
"facebook/xglm-564M": 7,
"google-bert/bert-base-multilingual-cased": 7,
"google/byt5-small": 29,
"google/gemma-2-2b": 7,
"gpt2": 7,
"meta-llama/Llama-3.2-1B": 7,
"microsoft/Phi-3-mini-4k-instruct": 7,
"mistralai/tekken": 7,
"tiktoken/gpt-4o": 7,
"tokenmonster/englishcode-32000-consistent-v1": 7
}
|
||
The metal with chemical sym. 'Fe' is
|
[
"lead",
"zinc",
"iron",
"gold"
] | 2 |
C
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
336-0.3
|
336.0
|
0.3
|
{
"CohereLabs/aya-expanse-8b": 0.9029386043548584,
"Qwen/Qwen3-8B": 0.9102910757064819,
"bigscience/bloom": 0.8799982070922852,
"common-pile/comma-v0.1-1t": 0.9248884916305542,
"facebook/xglm-564M": 0.8538807034492493,
"google-bert/bert-base-multilingual-cased": 0.8467434644699097,
"google/byt5-small": 0.9709159731864929,
"google/gemma-2-2b": 0.9017316699028015,
"gpt2": 0.8987561464309692,
"meta-llama/Llama-3.2-1B": 0.9013137221336365,
"microsoft/Phi-3-mini-4k-instruct": 0.8870991468429565,
"mistralai/tekken": 0.8897440433502197,
"tiktoken/gpt-4o": 0.913067102432251,
"tokenmonster/englishcode-32000-consistent-v1": 0.868256688117981
}
|
{
"CohereLabs/aya-expanse-8b": 0.14960479736328125,
"Qwen/Qwen3-8B": 0.10851742327213287,
"bigscience/bloom": 0.05709414184093475,
"common-pile/comma-v0.1-1t": 0.03906556963920593,
"facebook/xglm-564M": 0.06293395161628723,
"google-bert/bert-base-multilingual-cased": 0.03629113733768463,
"google/byt5-small": -0.034767430275678635,
"google/gemma-2-2b": 0.10764612257480621,
"gpt2": 0.08282459527254105,
"meta-llama/Llama-3.2-1B": 0.1575041115283966,
"microsoft/Phi-3-mini-4k-instruct": 0.050431989133358,
"mistralai/tekken": 0.10146505385637283,
"tiktoken/gpt-4o": 0.2269553542137146,
"tokenmonster/englishcode-32000-consistent-v1": -0.060447804629802704
}
|
{
"CohereLabs/aya-expanse-8b": 10,
"Qwen/Qwen3-8B": 10,
"bigscience/bloom": 10,
"common-pile/comma-v0.1-1t": 15,
"facebook/xglm-564M": 11,
"google-bert/bert-base-multilingual-cased": 11,
"google/byt5-small": 36,
"google/gemma-2-2b": 10,
"gpt2": 10,
"meta-llama/Llama-3.2-1B": 10,
"microsoft/Phi-3-mini-4k-instruct": 10,
"mistralai/tekken": 10,
"tiktoken/gpt-4o": 10,
"tokenmonster/englishcode-32000-consistent-v1": 10
}
|
||
The planet closest to the Sun in our solar system is
|
[
"♀",
"♂",
"☿",
"♁"
] | 2 |
C
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
338-0.4
|
338.0
|
0.4
|
{
"CohereLabs/aya-expanse-8b": 0.9999999403953552,
"Qwen/Qwen3-8B": 1,
"bigscience/bloom": 0.9999999403953552,
"common-pile/comma-v0.1-1t": 1.0000001192092896,
"facebook/xglm-564M": 0.9999999403953552,
"google-bert/bert-base-multilingual-cased": 0.9999997019767761,
"google/byt5-small": 1,
"google/gemma-2-2b": 0.9999999403953552,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 0.9999998807907104,
"microsoft/Phi-3-mini-4k-instruct": 0.9999999403953552,
"mistralai/tekken": 0.9999999403953552,
"tiktoken/gpt-4o": 1.0000001192092896,
"tokenmonster/englishcode-32000-consistent-v1": 0.9999998211860657
}
|
{
"CohereLabs/aya-expanse-8b": 1.0000003576278687,
"Qwen/Qwen3-8B": 1.0000004768371582,
"bigscience/bloom": 1.000000238418579,
"common-pile/comma-v0.1-1t": 1.0000004768371582,
"facebook/xglm-564M": 1.000000238418579,
"google-bert/bert-base-multilingual-cased": 1.0000004768371582,
"google/byt5-small": 1.0000004768371582,
"google/gemma-2-2b": 1.0000003576278687,
"gpt2": 1.0000004768371582,
"meta-llama/Llama-3.2-1B": 1.0000003576278687,
"microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687,
"mistralai/tekken": 1.0000004768371582,
"tiktoken/gpt-4o": 1.0000004768371582,
"tokenmonster/englishcode-32000-consistent-v1": 1.000000238418579
}
|
{
"CohereLabs/aya-expanse-8b": 11,
"Qwen/Qwen3-8B": 11,
"bigscience/bloom": 11,
"common-pile/comma-v0.1-1t": 14,
"facebook/xglm-564M": 12,
"google-bert/bert-base-multilingual-cased": 11,
"google/byt5-small": 52,
"google/gemma-2-2b": 11,
"gpt2": 11,
"meta-llama/Llama-3.2-1B": 11,
"microsoft/Phi-3-mini-4k-instruct": 11,
"mistralai/tekken": 11,
"tiktoken/gpt-4o": 11,
"tokenmonster/englishcode-32000-consistent-v1": 8
}
|
||
The largest planet in the Solar System is
|
[
"♁",
"♃",
"♄",
"♂"
] | 1 |
B
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
339-0.4
|
339.0
|
0.4
|
{
"CohereLabs/aya-expanse-8b": 0.9999998211860657,
"Qwen/Qwen3-8B": 0.9999998211860657,
"bigscience/bloom": 0.9999997615814209,
"common-pile/comma-v0.1-1t": 1.0000001192092896,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 0.9999999403953552,
"google/byt5-small": 1.000000238418579,
"google/gemma-2-2b": 1,
"gpt2": 0.9999996423721313,
"meta-llama/Llama-3.2-1B": 1,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 0.9999998211860657,
"tiktoken/gpt-4o": 0.9999998211860657,
"tokenmonster/englishcode-32000-consistent-v1": 0.9999998211860657
}
|
{
"CohereLabs/aya-expanse-8b": 1.0000003576278687,
"Qwen/Qwen3-8B": 1.0000004768371582,
"bigscience/bloom": 1.000000238418579,
"common-pile/comma-v0.1-1t": 1.0000004768371582,
"facebook/xglm-564M": 1.000000238418579,
"google-bert/bert-base-multilingual-cased": 1.0000004768371582,
"google/byt5-small": 1.0000004768371582,
"google/gemma-2-2b": 1.0000003576278687,
"gpt2": 1.0000004768371582,
"meta-llama/Llama-3.2-1B": 1.0000003576278687,
"microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687,
"mistralai/tekken": 1.0000004768371582,
"tiktoken/gpt-4o": 1.0000004768371582,
"tokenmonster/englishcode-32000-consistent-v1": 1.000000238418579
}
|
{
"CohereLabs/aya-expanse-8b": 8,
"Qwen/Qwen3-8B": 8,
"bigscience/bloom": 8,
"common-pile/comma-v0.1-1t": 9,
"facebook/xglm-564M": 8,
"google-bert/bert-base-multilingual-cased": 8,
"google/byt5-small": 41,
"google/gemma-2-2b": 8,
"gpt2": 8,
"meta-llama/Llama-3.2-1B": 8,
"microsoft/Phi-3-mini-4k-instruct": 9,
"mistralai/tekken": 8,
"tiktoken/gpt-4o": 8,
"tokenmonster/englishcode-32000-consistent-v1": 8
}
|
||
The process that allows plants to prod. their own food using sunlight is
|
[
"photosynthesis",
"respiration",
"digestion",
"fermentation"
] | 0 |
A
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
340-0.4
|
340.0
|
0.4
|
{
"CohereLabs/aya-expanse-8b": 0.9140673875808716,
"Qwen/Qwen3-8B": 0.9345059990882874,
"bigscience/bloom": 0.9172125458717346,
"common-pile/comma-v0.1-1t": 0.9256795048713684,
"facebook/xglm-564M": 0.9292510747909546,
"google-bert/bert-base-multilingual-cased": 0.8929300308227539,
"google/byt5-small": 0.993126630783081,
"google/gemma-2-2b": 0.9281373620033264,
"gpt2": 0.9144498705863953,
"meta-llama/Llama-3.2-1B": 0.9152953624725342,
"microsoft/Phi-3-mini-4k-instruct": 0.9282230138778687,
"mistralai/tekken": 0.917586088180542,
"tiktoken/gpt-4o": 0.9270635843276978,
"tokenmonster/englishcode-32000-consistent-v1": 0.8559166193008423
}
|
{
"CohereLabs/aya-expanse-8b": 0.0885179415345192,
"Qwen/Qwen3-8B": 0.19490374624729156,
"bigscience/bloom": 0.0557279996573925,
"common-pile/comma-v0.1-1t": 0.08006878197193146,
"facebook/xglm-564M": 0.11164054274559021,
"google-bert/bert-base-multilingual-cased": 0.0015388152096420527,
"google/byt5-small": -0.029237380251288414,
"google/gemma-2-2b": 0.14768964052200317,
"gpt2": 0.08957405388355255,
"meta-llama/Llama-3.2-1B": 0.0850566104054451,
"microsoft/Phi-3-mini-4k-instruct": 0.06736037880182266,
"mistralai/tekken": 0.07164368778467178,
"tiktoken/gpt-4o": 0.11323671042919159,
"tokenmonster/englishcode-32000-consistent-v1": -0.03460094332695007
}
|
{
"CohereLabs/aya-expanse-8b": 14,
"Qwen/Qwen3-8B": 14,
"bigscience/bloom": 14,
"common-pile/comma-v0.1-1t": 16,
"facebook/xglm-564M": 15,
"google-bert/bert-base-multilingual-cased": 16,
"google/byt5-small": 72,
"google/gemma-2-2b": 14,
"gpt2": 14,
"meta-llama/Llama-3.2-1B": 14,
"microsoft/Phi-3-mini-4k-instruct": 15,
"mistralai/tekken": 14,
"tiktoken/gpt-4o": 14,
"tokenmonster/englishcode-32000-consistent-v1": 12
}
|
||
The author who wrote the play "Romeo and Juliet" is
|
[
"W. Shakespeare",
"C. Dickens",
"M. Twain",
"J. Austen"
] | 0 |
A
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
341-0.2
|
341.0
|
0.2
|
{
"CohereLabs/aya-expanse-8b": 0.9999999403953552,
"Qwen/Qwen3-8B": 1.0000001192092896,
"bigscience/bloom": 1.0000001192092896,
"common-pile/comma-v0.1-1t": 1.0000001192092896,
"facebook/xglm-564M": 1.0000001192092896,
"google-bert/bert-base-multilingual-cased": 0.9999998211860657,
"google/byt5-small": 1.0000001192092896,
"google/gemma-2-2b": 1,
"gpt2": 1,
"meta-llama/Llama-3.2-1B": 0.9999999403953552,
"microsoft/Phi-3-mini-4k-instruct": 1,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 1.000000238418579,
"tokenmonster/englishcode-32000-consistent-v1": 1
}
|
{
"CohereLabs/aya-expanse-8b": 1.0000003576278687,
"Qwen/Qwen3-8B": 1.0000004768371582,
"bigscience/bloom": 1.000000238418579,
"common-pile/comma-v0.1-1t": 1.0000004768371582,
"facebook/xglm-564M": 1.000000238418579,
"google-bert/bert-base-multilingual-cased": 1.0000004768371582,
"google/byt5-small": 1.0000004768371582,
"google/gemma-2-2b": 1.0000003576278687,
"gpt2": 1.0000004768371582,
"meta-llama/Llama-3.2-1B": 1.0000003576278687,
"microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687,
"mistralai/tekken": 1.0000004768371582,
"tiktoken/gpt-4o": 1.0000004768371582,
"tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687
}
|
{
"CohereLabs/aya-expanse-8b": 12,
"Qwen/Qwen3-8B": 14,
"bigscience/bloom": 13,
"common-pile/comma-v0.1-1t": 17,
"facebook/xglm-564M": 14,
"google-bert/bert-base-multilingual-cased": 12,
"google/byt5-small": 51,
"google/gemma-2-2b": 12,
"gpt2": 14,
"meta-llama/Llama-3.2-1B": 14,
"microsoft/Phi-3-mini-4k-instruct": 15,
"mistralai/tekken": 13,
"tiktoken/gpt-4o": 13,
"tokenmonster/englishcode-32000-consistent-v1": 14
}
|
||
What plants need from the air to make food is
|
[
"N₂",
"H₂",
"He",
"CO₂"
] | 3 |
D
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
343-0.3
|
343.0
|
0.3
|
{
"CohereLabs/aya-expanse-8b": 1,
"Qwen/Qwen3-8B": 0.9999999403953552,
"bigscience/bloom": 1.0000001192092896,
"common-pile/comma-v0.1-1t": 0.9999998211860657,
"facebook/xglm-564M": 1,
"google-bert/bert-base-multilingual-cased": 0.9999998807907104,
"google/byt5-small": 1,
"google/gemma-2-2b": 0.9999997615814209,
"gpt2": 1.0000001192092896,
"meta-llama/Llama-3.2-1B": 1.0000001192092896,
"microsoft/Phi-3-mini-4k-instruct": 0.9999998807907104,
"mistralai/tekken": 1,
"tiktoken/gpt-4o": 0.9999998211860657,
"tokenmonster/englishcode-32000-consistent-v1": 0.9999998807907104
}
|
{
"CohereLabs/aya-expanse-8b": 1.0000003576278687,
"Qwen/Qwen3-8B": 1.0000004768371582,
"bigscience/bloom": 1.000000238418579,
"common-pile/comma-v0.1-1t": 1.0000004768371582,
"facebook/xglm-564M": 1.000000238418579,
"google-bert/bert-base-multilingual-cased": 1.0000004768371582,
"google/byt5-small": 1.0000004768371582,
"google/gemma-2-2b": 1.0000003576278687,
"gpt2": 1.0000004768371582,
"meta-llama/Llama-3.2-1B": 1.0000003576278687,
"microsoft/Phi-3-mini-4k-instruct": 1.0000003576278687,
"mistralai/tekken": 1.0000004768371582,
"tiktoken/gpt-4o": 1.0000004768371582,
"tokenmonster/englishcode-32000-consistent-v1": 1.0000003576278687
}
|
{
"CohereLabs/aya-expanse-8b": 10,
"Qwen/Qwen3-8B": 10,
"bigscience/bloom": 10,
"common-pile/comma-v0.1-1t": 12,
"facebook/xglm-564M": 10,
"google-bert/bert-base-multilingual-cased": 10,
"google/byt5-small": 45,
"google/gemma-2-2b": 10,
"gpt2": 10,
"meta-llama/Llama-3.2-1B": 10,
"microsoft/Phi-3-mini-4k-instruct": 10,
"mistralai/tekken": 10,
"tiktoken/gpt-4o": 10,
"tokenmonster/englishcode-32000-consistent-v1": 8
}
|
||
In "Can you pls. book a flight to Paris?", the person wants to
|
[
"go shopping",
"file a complaint",
"cancel reservation",
"make a booking"
] | 3 |
D
|
test
|
Abbreviations
|
Script / Orthography
|
eng_Latn
|
344-0.4
|
344.0
|
0.4
|
{
"CohereLabs/aya-expanse-8b": 0.8967032432556152,
"Qwen/Qwen3-8B": 0.9513365030288696,
"bigscience/bloom": 0.9068660140037537,
"common-pile/comma-v0.1-1t": 0.958304226398468,
"facebook/xglm-564M": 0.9203377366065979,
"google-bert/bert-base-multilingual-cased": 0.9147680997848511,
"google/byt5-small": 0.9851800203323364,
"google/gemma-2-2b": 0.9473385214805603,
"gpt2": 0.8919826149940491,
"meta-llama/Llama-3.2-1B": 0.9390807747840881,
"microsoft/Phi-3-mini-4k-instruct": 0.9023863077163696,
"mistralai/tekken": 0.884173572063446,
"tiktoken/gpt-4o": 0.9475116729736328,
"tokenmonster/englishcode-32000-consistent-v1": 0.8966159820556641
}
|
{
"CohereLabs/aya-expanse-8b": -0.009711533784866333,
"Qwen/Qwen3-8B": 0.1350504606962204,
"bigscience/bloom": 0.01425527036190033,
"common-pile/comma-v0.1-1t": 0.13568083941936493,
"facebook/xglm-564M": 0.06789200007915497,
"google-bert/bert-base-multilingual-cased": 0.08616456389427185,
"google/byt5-small": 0.30707496404647827,
"google/gemma-2-2b": 0.1704685240983963,
"gpt2": 0.0131788719445467,
"meta-llama/Llama-3.2-1B": 0.140251025557518,
"microsoft/Phi-3-mini-4k-instruct": 0.04106447473168373,
"mistralai/tekken": 0.01378481462597847,
"tiktoken/gpt-4o": 0.10346754640340805,
"tokenmonster/englishcode-32000-consistent-v1": 0.07150956243276596
}
|
{
"CohereLabs/aya-expanse-8b": 17,
"Qwen/Qwen3-8B": 16,
"bigscience/bloom": 18,
"common-pile/comma-v0.1-1t": 20,
"facebook/xglm-564M": 18,
"google-bert/bert-base-multilingual-cased": 18,
"google/byt5-small": 62,
"google/gemma-2-2b": 16,
"gpt2": 17,
"meta-llama/Llama-3.2-1B": 16,
"microsoft/Phi-3-mini-4k-instruct": 18,
"mistralai/tekken": 17,
"tiktoken/gpt-4o": 16,
"tokenmonster/englishcode-32000-consistent-v1": 14
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.