tiny_qa_benchmark / README.md
vincentkoc's picture
Update README.md
7784faa verified
---
license: apache-2.0
language:
- en
pretty_name: Tiny QA Benchmark (Original EN Core for TQB++)
size_categories:
- n<1K
tags:
- question-answering
- evaluation
- benchmark
- toy-dataset
- tqb++-core
task_categories:
- question-answering
task_ids:
- extractive-qa
- closed-book-qa
arxiv: 2505.12058
datasets:
- vincentkoc/tiny_qa_benchmark_pp
---
# Tiny QA Benchmark (Original English Core for TQB++)
**This dataset (`vincentkoc/tiny_qa_benchmark`) is the original 52-item English Question-Answering set. It now serves as the immutable "gold standard" core for the expanded [Tiny QA Benchmark++ (TQB++)](https://huggingface.co/datasets/vincentkoc/tiny_qa_benchmark_pp) project.**
The TQB++ project builds upon this core dataset by introducing a powerful synthetic generation toolkit, pre-built multilingual datasets, and a comprehensive framework for rapid LLM smoke testing.
**For the full TQB++ toolkit, the latest research paper, multilingual datasets, and the synthetic generator, please visit:**
* **TQB++ Hugging Face Dataset Collection & Toolkit:** [vincentkoc/tiny_qa_benchmark_pp](https://huggingface.co/datasets/vincentkoc/tiny_qa_benchmark_pp)
* **TQB++ GitHub Repository (Code, Paper & Toolkit):** [vincentkoc/tiny_qa_benchmark_pp](https://github.com/vincentkoc/tiny_qa_benchmark_pp)
This original dataset (`vincentkoc/tiny_qa_benchmark`) contains 52 hand-crafted general-knowledge QA pairs covering geography, history, math, science, literature, and more. It remains ideal for quick sanity checks, pipeline smoke-tests, and as a foundational component of TQB++. Each example includes:
- **text**: the question prompt
- **label**: the “gold” answer
- **metadata.context**: a one-sentence fact
- **tags**: additional annotations (`category`, `difficulty`)
It’s intentionally tiny (<100 KB) so you can iterate on data loading, evaluation scripts, or CI steps in under a second using these specific 52 items.
## Supported Tasks and Formats (for this core dataset)
- **Tasks**:
- Extractive QA
- Generative QA
- **Format**: JSON
- **Splits**:
- `train` (all 52 examples)
## Languages (for this core dataset)
- English (`en`)
## Dataset Structure
### Data Fields
Each example in `data/train.json` (as loaded by `datasets`) has:
| field | type | description |
|---------------------|--------|----------------------------------------------|
| `text` | string | The question prompt. |
| `label` | string | The correct answer. |
| `metadata` | object | Additional info. |
| `metadata.context` | string | A one-sentence fact supporting the answer. |
| `tags.category` | string | Broad question category (e.g. `geography`). |
| `tags.difficulty` | string | Rough difficulty level (e.g. `easy`). |
## Data Example
```json
[
{
"text": "What is the capital of France?",
"label": "Paris",
"metadata": {
"context": "France is a country in Europe. Its capital is Paris."
},
"tags": {
"category": "geography",
"difficulty": "easy"
}
},
{
"text": "What is 2 + 2?",
"label": "4",
"metadata": {
"context": "Basic arithmetic: 2 + 2 equals 4."
},
"tags": {
"category": "math",
"difficulty": "easy"
}
}
]
```
*(Note: The actual file on the Hub might be a `.jsonl` file where each line is a JSON object, but `load_dataset` handles this.)*
## Data Splits
Only one split for this core dataset:
- **train**: 52 examples, used for development, quick evaluation, and as the TQB++ core.
## Data Creation
### Curation Rationale
The "Tiny QA Benchmark" (this 52-item set) was originally created to:
1. Smoke-test QA pipelines (loading, preprocessing, evaluation).
2. Demo Hugging Face Datasets integration in tutorials.
3. Verify model–eval loops run without downloading large corpora.
4. **Serve as the immutable "gold standard" English core for the [Tiny QA Benchmark++ (TQB++)](https://huggingface.co/datasets/vincentkoc/tiny_qa_benchmark_pp) project.**
### Source Data
Hand-crafted by the dataset creator from well-known, public-domain facts.
It was initially developed as a dataset for sample projects to demonstrate [Opik](https://github.com/comet-ml/opik/) and now forms the foundational English core of TQB++.
### Annotations
Self-annotated. Each `metadata.context` and `tags` field is manually created for these 52 items.
## Usage
Load this specific 52-item core dataset with:
```python
from datasets import load_dataset
ds = load_dataset("vincentkoc/tiny_qa_benchmark")
print(ds["train"][0])
# Expected output:
# {
# "text": "What is the capital of France?",
# "label": "Paris",
# "metadata": {
# "context": "France is a country in Europe. Its capital is Paris."
# },
# "tags": {
# "category": "geography",
# "difficulty": "easy"
# }
# }
```
For accessing the full TQB++ suite, including multilingual packs and the synthetic generator, refer to the [TQB++ Hugging Face Dataset Collection](https://huggingface.co/datasets/vincentkoc/tiny_qa_benchmark_pp).
## Considerations for Use
* **Immutable Core for TQB++:** This dataset is the stable, hand-curated English core of the TQB++ project. Its 52 items are not intended to change.
* **Not a Comprehensive Benchmark (on its own):** While excellent for quick checks, these 52 items are too few for statistically significant model ranking. For broader evaluation, use in conjunction with the TQB++ synthetic generator and its multilingual capabilities found at [vincentkoc/tiny_qa_benchmark_pp](https://huggingface.co/datasets/vincentkoc/tiny_qa_benchmark_pp).
* **Do Not Train:** Primarily intended for evaluation, smoke-tests, or demos.
* **No Sensitive Data:** All facts are public domain.
## Licensing
Apache-2.0. See the `LICENSE` file in the [TQB++ GitHub repository](https://github.com/vincentkoc/tiny_qa_benchmark_pp) for details (as this dataset is now part of that larger project).
## Citation
If you use this specific 52-item core English dataset, please cite it. You can use the following BibTeX entry, which has been updated to reflect its role:
```bibtex
@misc{koctinyqabenchmark_original_core,
author = { Vincent Koc },
title = { Tiny QA Benchmark (Original 52-item English Core for TQB++) },
year = 2025,
url = { https://huggingface.co/datasets/vincentkoc/tiny_qa_benchmark },
doi = { 10.57967/hf/5417 },
publisher = { Hugging Face }
}
```
For the complete **Tiny QA Benchmark++ (TQB++)** project (which includes this core set, the synthetic generator, multilingual packs, and the associated research paper), please refer to and cite the TQB++ project directly:
```bibtex
@misc{koctinyqabenchmark_pp_dataset,
author = {Vincent Koc},
title = {Tiny QA Benchmark++ (TQB++) Datasets and Toolkit},
year = {2025},
publisher = {Hugging Face & GitHub},
doi = {10.57967/hf/5531}, /* DOI for the TQB++ collection */
howpublished = {\\url{https://huggingface.co/datasets/vincentkoc/tiny_qa_benchmark_pp}},
note = {See also: \\url{https://github.com/vincentkoc/tiny_qa_benchmark_pp}}
}
```