Commit
·
f52f478
1
Parent(s):
bbad117
update script and data card
Browse files
P3.py
CHANGED
|
@@ -12,7 +12,7 @@
|
|
| 12 |
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 13 |
# See the License for the specific language governing permissions and
|
| 14 |
# limitations under the License.
|
| 15 |
-
"""P3"""
|
| 16 |
|
| 17 |
|
| 18 |
import datasets
|
|
@@ -27,7 +27,7 @@ _CITATION = """\
|
|
| 27 |
TODO"""
|
| 28 |
|
| 29 |
_DESCRIPTION = """\
|
| 30 |
-
P3 is a collection of prompted English datasets covering a diverse set of NLP tasks. A prompt is the combination of an input template and a target template. The templates are functions mapping a data example into natural language for the input and target sequences. For example, in the case of an NLI dataset, the data example would include fields for *Premise, Hypothesis, Label*. An input template would be *If {Premise} is true, is it also true that {Hypothesis}?*, whereas a target template can be defined with the label choices *Choices[label]*. Here *Choices* is prompt-specific metadata that consists of the options *yes, maybe, no* corresponding to *label* being entailment (0), neutral (1) or contradiction (2).
|
| 31 |
|
| 32 |
Prompts are collected using [Promptsource](https://github.com/bigscience-workshop/promptsource), an interface to interactively write prompts on datasets, and collect prompt-specific metadata such as evaluation metrics. As of October 13th, there are 2'000 prompts collected for 270+ data(sub)sets. The collection of prompts is publicly available on [Promptsource](https://github.com/bigscience-workshop/promptsource).
|
| 33 |
|
|
@@ -83,8 +83,6 @@ def find_task_splits_and_features():
|
|
| 83 |
"""Find the available tasks under ./data and their available splits and features."""
|
| 84 |
task_and_their_splits = defaultdict(dict)
|
| 85 |
for stats in glob.glob(f"{_DATA_PATH}/*/stats.*.json"):
|
| 86 |
-
if "anli" not in stats:
|
| 87 |
-
continue
|
| 88 |
folder_path = os.path.dirname(stats)
|
| 89 |
task_name = folder_path.split("/")[-1]
|
| 90 |
split_name = os.path.basename(stats).split(".")[1]
|
|
|
|
| 12 |
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 13 |
# See the License for the specific language governing permissions and
|
| 14 |
# limitations under the License.
|
| 15 |
+
"""P3 (Public Pool of Prompts)"""
|
| 16 |
|
| 17 |
|
| 18 |
import datasets
|
|
|
|
| 27 |
TODO"""
|
| 28 |
|
| 29 |
_DESCRIPTION = """\
|
| 30 |
+
P3 (Pubic Pool of Prompts)is a collection of prompted English datasets covering a diverse set of NLP tasks. A prompt is the combination of an input template and a target template. The templates are functions mapping a data example into natural language for the input and target sequences. For example, in the case of an NLI dataset, the data example would include fields for *Premise, Hypothesis, Label*. An input template would be *If {Premise} is true, is it also true that {Hypothesis}?*, whereas a target template can be defined with the label choices *Choices[label]*. Here *Choices* is prompt-specific metadata that consists of the options *yes, maybe, no* corresponding to *label* being entailment (0), neutral (1) or contradiction (2).
|
| 31 |
|
| 32 |
Prompts are collected using [Promptsource](https://github.com/bigscience-workshop/promptsource), an interface to interactively write prompts on datasets, and collect prompt-specific metadata such as evaluation metrics. As of October 13th, there are 2'000 prompts collected for 270+ data(sub)sets. The collection of prompts is publicly available on [Promptsource](https://github.com/bigscience-workshop/promptsource).
|
| 33 |
|
|
|
|
| 83 |
"""Find the available tasks under ./data and their available splits and features."""
|
| 84 |
task_and_their_splits = defaultdict(dict)
|
| 85 |
for stats in glob.glob(f"{_DATA_PATH}/*/stats.*.json"):
|
|
|
|
|
|
|
| 86 |
folder_path = os.path.dirname(stats)
|
| 87 |
task_name = folder_path.split("/")[-1]
|
| 88 |
split_name = os.path.basename(stats).split(".")[1]
|
README.md
CHANGED
|
@@ -32,13 +32,7 @@ task_categories:
|
|
| 32 |
- [Curation Rationale](#curation-rationale)
|
| 33 |
- [Source Data](#source-data)
|
| 34 |
- [Annotations](#annotations)
|
| 35 |
-
- [Personal and Sensitive Information](#personal-and-sensitive-information)
|
| 36 |
-
- [Considerations for Using the Data](#considerations-for-using-the-data)
|
| 37 |
-
- [Social Impact of Dataset](#social-impact-of-dataset)
|
| 38 |
-
- [Discussion of Biases](#discussion-of-biases)
|
| 39 |
-
- [Other Known Limitations](#other-known-limitations)
|
| 40 |
- [Additional Information](#additional-information)
|
| 41 |
-
- [Dataset Curators](#dataset-curators)
|
| 42 |
- [Licensing Information](#licensing-information)
|
| 43 |
- [Citation Information](#citation-information)
|
| 44 |
- [Contributions](#contributions)
|
|
@@ -183,6 +177,8 @@ The main annotation guideline was that prompts needed to be grammatical and unde
|
|
| 183 |
|
| 184 |
The full annotation given to the contributors can be found [here](https://github.com/bigscience-workshop/promptsource/blob/main/CONTRIBUTING.md). *Note to self: the link is currently being updated with the)
|
| 185 |
|
|
|
|
|
|
|
| 186 |
### Licensing Information
|
| 187 |
|
| 188 |
The dataset is released under Apache 2.0.
|
|
|
|
| 32 |
- [Curation Rationale](#curation-rationale)
|
| 33 |
- [Source Data](#source-data)
|
| 34 |
- [Annotations](#annotations)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 35 |
- [Additional Information](#additional-information)
|
|
|
|
| 36 |
- [Licensing Information](#licensing-information)
|
| 37 |
- [Citation Information](#citation-information)
|
| 38 |
- [Contributions](#contributions)
|
|
|
|
| 177 |
|
| 178 |
The full annotation given to the contributors can be found [here](https://github.com/bigscience-workshop/promptsource/blob/main/CONTRIBUTING.md). *Note to self: the link is currently being updated with the)
|
| 179 |
|
| 180 |
+
## Additional Information
|
| 181 |
+
|
| 182 |
### Licensing Information
|
| 183 |
|
| 184 |
The dataset is released under Apache 2.0.
|