Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
pandas
License:
SaylorTwift HF Staff commited on
Commit
d7bc9b8
·
verified ·
1 Parent(s): 0a7edb8

Add 'college_computer_science' config data files

Browse files
README.md CHANGED
@@ -274,6 +274,8 @@ dataset_info:
274
  features:
275
  - name: question
276
  dtype: string
 
 
277
  - name: choices
278
  sequence: string
279
  - name: answer
@@ -286,19 +288,19 @@ dataset_info:
286
  '3': D
287
  splits:
288
  - name: auxiliary_train
289
- num_bytes: 160601377
290
  num_examples: 99842
291
  - name: test
292
- num_bytes: 42641
293
  num_examples: 100
294
  - name: validation
295
- num_bytes: 4663
296
  num_examples: 11
297
  - name: dev
298
- num_bytes: 2765
299
  num_examples: 5
300
- download_size: 166184960
301
- dataset_size: 160651446
302
  - config_name: college_mathematics
303
  features:
304
  - name: question
@@ -1801,6 +1803,16 @@ configs:
1801
  path: college_chemistry/validation-*
1802
  - split: dev
1803
  path: college_chemistry/dev-*
 
 
 
 
 
 
 
 
 
 
1804
  ---
1805
 
1806
  # Dataset Card for MMLU
 
274
  features:
275
  - name: question
276
  dtype: string
277
+ - name: subject
278
+ dtype: string
279
  - name: choices
280
  sequence: string
281
  - name: answer
 
288
  '3': D
289
  splits:
290
  - name: auxiliary_train
291
+ num_bytes: 161000625
292
  num_examples: 99842
293
  - name: test
294
+ num_bytes: 45429
295
  num_examples: 100
296
  - name: validation
297
+ num_bytes: 4959
298
  num_examples: 11
299
  - name: dev
300
+ num_bytes: 2893
301
  num_examples: 5
302
+ download_size: 47201966
303
+ dataset_size: 161053906
304
  - config_name: college_mathematics
305
  features:
306
  - name: question
 
1803
  path: college_chemistry/validation-*
1804
  - split: dev
1805
  path: college_chemistry/dev-*
1806
+ - config_name: college_computer_science
1807
+ data_files:
1808
+ - split: auxiliary_train
1809
+ path: college_computer_science/auxiliary_train-*
1810
+ - split: test
1811
+ path: college_computer_science/test-*
1812
+ - split: validation
1813
+ path: college_computer_science/validation-*
1814
+ - split: dev
1815
+ path: college_computer_science/dev-*
1816
  ---
1817
 
1818
  # Dataset Card for MMLU
college_computer_science/auxiliary_train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2782fc860f57d9345a9233ab04f494b0af5ae85b893a27853f7014b14a3bd07
3
+ size 47163955
college_computer_science/dev-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4882d0d971ddb2842812ae25a260097c4a4cb719720e5e8dd91e33d84d863185
3
+ size 5614
college_computer_science/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8685a45ae7f9224840bd39704512f7389e671ab5a40e5fbabc93cbba27013b5
3
+ size 26811
college_computer_science/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7317ea44dfcfc8b66dddfe53618421add2e507d1a6736f85b12449828de0c63
3
+ size 5586
dataset_infos.json CHANGED
@@ -497,39 +497,34 @@
497
  "features": {
498
  "question": {
499
  "dtype": "string",
500
- "id": null,
 
 
 
501
  "_type": "Value"
502
  },
503
  "choices": {
504
  "feature": {
505
  "dtype": "string",
506
- "id": null,
507
  "_type": "Value"
508
  },
509
- "length": -1,
510
- "id": null,
511
  "_type": "Sequence"
512
  },
513
  "answer": {
514
- "num_classes": 4,
515
  "names": [
516
  "A",
517
  "B",
518
  "C",
519
  "D"
520
  ],
521
- "id": null,
522
  "_type": "ClassLabel"
523
  }
524
  },
525
- "post_processed": null,
526
- "supervised_keys": null,
527
- "task_templates": null,
528
- "builder_name": "mmlu",
529
  "config_name": "college_computer_science",
530
  "version": {
531
  "version_str": "1.0.0",
532
- "description": null,
533
  "major": 1,
534
  "minor": 0,
535
  "patch": 0
@@ -537,39 +532,32 @@
537
  "splits": {
538
  "auxiliary_train": {
539
  "name": "auxiliary_train",
540
- "num_bytes": 160601257,
541
  "num_examples": 99842,
542
- "dataset_name": "mmlu"
543
  },
544
  "test": {
545
  "name": "test",
546
- "num_bytes": 42629,
547
  "num_examples": 100,
548
- "dataset_name": "mmlu"
549
  },
550
  "validation": {
551
  "name": "validation",
552
- "num_bytes": 4651,
553
  "num_examples": 11,
554
- "dataset_name": "mmlu"
555
  },
556
  "dev": {
557
  "name": "dev",
558
- "num_bytes": 2753,
559
  "num_examples": 5,
560
- "dataset_name": "mmlu"
561
- }
562
- },
563
- "download_checksums": {
564
- "data.tar": {
565
- "num_bytes": 166184960,
566
- "checksum": "bec563ba4bac1d6aaf04141cd7d1605d7a5ca833e38f994051e818489592989b"
567
  }
568
  },
569
- "download_size": 166184960,
570
- "post_processing_size": null,
571
- "dataset_size": 160651290,
572
- "size_in_bytes": 326836250
573
  },
574
  "college_mathematics": {
575
  "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n",
 
497
  "features": {
498
  "question": {
499
  "dtype": "string",
500
+ "_type": "Value"
501
+ },
502
+ "subject": {
503
+ "dtype": "string",
504
  "_type": "Value"
505
  },
506
  "choices": {
507
  "feature": {
508
  "dtype": "string",
 
509
  "_type": "Value"
510
  },
 
 
511
  "_type": "Sequence"
512
  },
513
  "answer": {
 
514
  "names": [
515
  "A",
516
  "B",
517
  "C",
518
  "D"
519
  ],
 
520
  "_type": "ClassLabel"
521
  }
522
  },
523
+ "builder_name": "parquet",
524
+ "dataset_name": "mmlu",
 
 
525
  "config_name": "college_computer_science",
526
  "version": {
527
  "version_str": "1.0.0",
 
528
  "major": 1,
529
  "minor": 0,
530
  "patch": 0
 
532
  "splits": {
533
  "auxiliary_train": {
534
  "name": "auxiliary_train",
535
+ "num_bytes": 161000625,
536
  "num_examples": 99842,
537
+ "dataset_name": null
538
  },
539
  "test": {
540
  "name": "test",
541
+ "num_bytes": 45429,
542
  "num_examples": 100,
543
+ "dataset_name": null
544
  },
545
  "validation": {
546
  "name": "validation",
547
+ "num_bytes": 4959,
548
  "num_examples": 11,
549
+ "dataset_name": null
550
  },
551
  "dev": {
552
  "name": "dev",
553
+ "num_bytes": 2893,
554
  "num_examples": 5,
555
+ "dataset_name": null
 
 
 
 
 
 
556
  }
557
  },
558
+ "download_size": 47201966,
559
+ "dataset_size": 161053906,
560
+ "size_in_bytes": 208255872
 
561
  },
562
  "college_mathematics": {
563
  "description": "This is a massive multitask test consisting of multiple-choice questions from various branches of knowledge, covering 57 tasks including elementary mathematics, US history, computer science, law, and more.\n",