import json import datasets _DESCRIPTION = """ SciFact A dataset of expert-written scientific claims paired with evidence-containing abstracts and annotated with labels and rationales. """ _CITATION = """ @InProceedings{Wadden2020FactOF, author = {David Wadden, Shanchuan Lin, Kyle Lo, Lucy Lu Wang, Madeleine van Zuylen, Arman Cohan, Hannaneh Hajishirzi}, title = {Fact or Fiction: Verifying Scientific Claims}, booktitle = {EMNLP}, year = 2020, } """ _DOWNLOAD_URL = "https://testerstories.com/files/ai_learn/data.tar.gz" class ScifactConfig(datasets.BuilderConfig): def __init__(self, **kwargs): super(ScifactConfig, self).__init__( version=datasets.Version("1.0.0", ""), **kwargs ) class Scifact(datasets.GeneratorBasedBuilder): VERSION = datasets.Version("0.1.0") BUILDER_CONFIGS = [ ScifactConfig(name="corpus", description="The corpus of evidence documents"), ScifactConfig( name="claims", description="The claims are split into train, test, dev" ), ] def _info(self): if self.config.name == "corpus": features = { "doc_id": datasets.Value("int32"), "title": datasets.Value("string"), "abstract": datasets.features.Sequence(datasets.Value("string")), "structured": datasets.Value("bool"), } else: features = { "id": datasets.Value("int32"), "claim": datasets.Value("string"), "evidence_doc_id": datasets.Value("string"), "evidence_label": datasets.Value("string"), "evidence_sentences": datasets.features.Sequence( datasets.Value("int32") ), "cited_doc_ids": datasets.features.Sequence(datasets.Value("int32")), } return datasets.DatasetInfo( description=_DESCRIPTION, features=datasets.Features(features), supervised_keys=None, homepage="https://scifact.apps.allenai.org/", citation=_CITATION, ) def _split_generators(self, dl_manager): archive = dl_manager.download(_DOWNLOAD_URL) if self.config.name == "corpus": return [ datasets.SplitGenerator( name=datasets.Split.TRAIN, gen_kwargs={ "filepath": "data/corpus.jsonl", "split": "train", "files": dl_manager.iter_archive(archive), }, ), ] else: return [ datasets.SplitGenerator( name=datasets.Split.TRAIN, gen_kwargs={ "filepath": "data/claims_train.jsonl", "split": "train", "files": dl_manager.iter_archive(archive), }, ), datasets.SplitGenerator( name=datasets.Split.TEST, gen_kwargs={ "filepath": "data/claims_test.jsonl", "split": "test", "files": dl_manager.iter_archive(archive), }, ), datasets.SplitGenerator( name=datasets.Split.VALIDATION, gen_kwargs={ "filepath": "data/claims_dev.jsonl", "split": "dev", "files": dl_manager.iter_archive(archive), }, ), ] def _generate_examples(self, filepath, split, files): for path, f in files: if path == filepath: for id_, row in enumerate(f): data = json.loads(row.decode("utf-8")) if self.config.name == "corpus": yield id_, { "doc_id": int(data["doc_id"]), "title": data["title"], "abstract": data["abstract"], "structured": data["structured"], } else: if split == "test": yield id_, { "id": data["id"], "claim": data["claim"], "evidence_doc_id": "", "evidence_label": "", "evidence_sentences": [], "cited_doc_ids": [], } else: evidences = data["evidence"] if evidences: for id1, doc_id in enumerate(evidences): for id2, evidence in enumerate(evidences[doc_id]): yield str(id_) + "_" + str(id1) + "_" + str( id2 ), { "id": data["id"], "claim": data["claim"], "evidence_doc_id": doc_id, "evidence_label": evidence["label"], "evidence_sentences": evidence["sentences"], "cited_doc_ids": data.get( "cited_doc_ids", [] ), } else: yield id_, { "id": data["id"], "claim": data["claim"], "evidence_doc_id": "", "evidence_label": "", "evidence_sentences": [], "cited_doc_ids": data.get("cited_doc_ids", []), } break