|
import os |
|
|
|
import datasets |
|
import numpy as np |
|
import pandas as pd |
|
|
|
from .bigbiohub import text_features |
|
from .bigbiohub import kb_features |
|
from .bigbiohub import BigBioConfig |
|
from .bigbiohub import Tasks |
|
|
|
_LANGUAGES = ['French'] |
|
_PUBMED = False |
|
_LOCAL = True |
|
_CITATION = """\ |
|
@misc{dalloux, title={Datasets – Clément Dalloux}, url={http://clementdalloux.fr/?page_id=28}, journal={Clément Dalloux}, author={Dalloux, Clément}} |
|
""" |
|
|
|
_DATASETNAME = "essai" |
|
_DISPLAYNAME = "ESSAI" |
|
|
|
_DESCRIPTION = """\ |
|
We manually annotated two corpora from the biomedical field. The ESSAI corpus \ |
|
contains clinical trial protocols in French. They were mainly obtained from the \ |
|
National Cancer Institute The typical protocol consists of two parts: the \ |
|
summary of the trial, which indicates the purpose of the trial and the methods \ |
|
applied; and a detailed description of the trial with the inclusion and \ |
|
exclusion criteria. The CAS corpus contains clinical cases published in \ |
|
scientific literature and training material. They are published in different \ |
|
journals from French-speaking countries (France, Belgium, Switzerland, Canada, \ |
|
African countries, tropical countries) and are related to various medical \ |
|
specialties (cardiology, urology, oncology, obstetrics, pulmonology, \ |
|
gastro-enterology). The purpose of clinical cases is to describe clinical \ |
|
situations of patients. Hence, their content is close to the content of clinical \ |
|
narratives (description of diagnoses, treatments or procedures, evolution, \ |
|
family history, expected audience, etc.). In clinical cases, the negation is \ |
|
frequently used for describing the patient signs, symptoms, and diagnosis. \ |
|
Speculation is present as well but less frequently. |
|
|
|
This version only contain the annotated ESSAI corpus |
|
""" |
|
|
|
_HOMEPAGE = "https://clementdalloux.fr/?page_id=28" |
|
|
|
_LICENSE = 'Data User Agreement' |
|
|
|
_URLS = { |
|
"essai_source": "", |
|
"essai_bigbio_text": "", |
|
"essai_bigbio_kb": "", |
|
} |
|
|
|
_SOURCE_VERSION = "1.0.0" |
|
_BIGBIO_VERSION = "1.0.0" |
|
|
|
_SUPPORTED_TASKS = [Tasks.TEXT_CLASSIFICATION] |
|
|
|
|
|
class ESSAI(datasets.GeneratorBasedBuilder): |
|
SOURCE_VERSION = datasets.Version(_SOURCE_VERSION) |
|
BIGBIO_VERSION = datasets.Version(_BIGBIO_VERSION) |
|
|
|
DEFAULT_CONFIG_NAME = "essai_source" |
|
|
|
BUILDER_CONFIGS = [ |
|
BigBioConfig( |
|
name="essai_source", |
|
version=SOURCE_VERSION, |
|
description="ESSAI source schema", |
|
schema="source", |
|
subset_id="essai", |
|
), |
|
BigBioConfig( |
|
name="essai_bigbio_text", |
|
version=BIGBIO_VERSION, |
|
description="ESSAI simplified BigBio schema for negation/speculation classification", |
|
schema="bigbio_text", |
|
subset_id="essai", |
|
), |
|
BigBioConfig( |
|
name="essai_bigbio_kb", |
|
version=BIGBIO_VERSION, |
|
description="ESSAI simplified BigBio schema for part-of-speech-tagging", |
|
schema="bigbio_kb", |
|
subset_id="essai", |
|
), |
|
] |
|
|
|
def _info(self): |
|
if self.config.schema == "source": |
|
features = datasets.Features( |
|
{ |
|
"id": datasets.Value("string"), |
|
"document_id": datasets.Value("string"), |
|
"text": [datasets.Value("string")], |
|
"lemmas": [datasets.Value("string")], |
|
"POS_tags": [datasets.Value("string")], |
|
"labels": [datasets.Value("string")], |
|
} |
|
) |
|
elif self.config.schema == "bigbio_text": |
|
features = text_features |
|
elif self.config.schema == "bigbio_kb": |
|
features = kb_features |
|
|
|
return datasets.DatasetInfo( |
|
description=_DESCRIPTION, |
|
features=features, |
|
supervised_keys=None, |
|
homepage=_HOMEPAGE, |
|
license=str(_LICENSE), |
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
if self.config.data_dir is None: |
|
raise ValueError( |
|
"This is a local dataset. Please pass the data_dir kwarg to load_dataset." |
|
) |
|
else: |
|
data_dir = self.config.data_dir |
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
gen_kwargs={"datadir": data_dir}, |
|
), |
|
] |
|
|
|
def _generate_examples(self, datadir): |
|
key = 0 |
|
for file in ["ESSAI_neg.txt", "ESSAI_spec.txt"]: |
|
filepath = os.path.join(datadir, file) |
|
label = "negation" if "neg" in file else "speculation" |
|
id_docs = [] |
|
id_words = [] |
|
words = [] |
|
lemmas = [] |
|
POS_tags = [] |
|
|
|
with open(filepath) as f: |
|
for line in f.readlines(): |
|
line_content = line.split("\t") |
|
if len(line_content) > 1: |
|
id_docs.append(line_content[0]) |
|
id_words.append(line_content[1]) |
|
words.append(line_content[2]) |
|
lemmas.append(line_content[3]) |
|
POS_tags.append(line_content[4]) |
|
|
|
dic = { |
|
"id_docs": np.array(list(map(int, id_docs))), |
|
"id_words": id_words, |
|
"words": words, |
|
"lemmas": lemmas, |
|
"POS_tags": POS_tags, |
|
} |
|
if self.config.schema == "source": |
|
for doc_id in set(dic["id_docs"]): |
|
idces = np.argwhere(dic["id_docs"] == doc_id)[:, 0] |
|
text = [dic["words"][id] for id in idces] |
|
text_lemmas = [dic["lemmas"][id] for id in idces] |
|
POS_tags_ = [dic["POS_tags"][id] for id in idces] |
|
yield key, { |
|
"id": key, |
|
"document_id": doc_id, |
|
"text": text, |
|
"lemmas": text_lemmas, |
|
"POS_tags": POS_tags_, |
|
"labels": [label], |
|
} |
|
key += 1 |
|
elif self.config.schema == "bigbio_text": |
|
for doc_id in set(dic["id_docs"]): |
|
idces = np.argwhere(dic["id_docs"] == doc_id)[:, 0] |
|
text = " ".join([dic["words"][id] for id in idces]) |
|
yield key, { |
|
"id": key, |
|
"document_id": doc_id, |
|
"text": text, |
|
"labels": [label], |
|
} |
|
key += 1 |
|
elif self.config.schema == "bigbio_kb": |
|
for doc_id in set(dic["id_docs"]): |
|
idces = np.argwhere(dic["id_docs"] == doc_id)[:, 0] |
|
text = [dic["words"][id] for id in idces] |
|
POS_tags_ = [dic["POS_tags"][id] for id in idces] |
|
|
|
data = { |
|
"id": str(key), |
|
"document_id": doc_id, |
|
"passages": [], |
|
"entities": [], |
|
"relations": [], |
|
"events": [], |
|
"coreferences": [], |
|
} |
|
key += 1 |
|
|
|
data["passages"] = [ |
|
{ |
|
"id": str(key + i), |
|
"type": "sentence", |
|
"text": [text[i]], |
|
"offsets": [[i, i + 1]], |
|
} |
|
for i in range(len(text)) |
|
] |
|
key += len(text) |
|
|
|
for i in range(len(text)): |
|
entity = { |
|
"id": key, |
|
"type": "POS_tag", |
|
"text": [POS_tags_[i]], |
|
"offsets": [[i, i + 1]], |
|
"normalized": [], |
|
} |
|
data["entities"].append(entity) |
|
key += 1 |
|
|
|
yield key, data |
|
|