The dataset viewer is not available for this dataset.
Error code: ConfigNamesError Exception: ImportError Message: To be able to use universalner/universal_ner, you need to install the following dependency: conllu. Please install it using 'pip install conllu' for instance. Traceback: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/dataset/config_names.py", line 66, in compute_config_names_response config_names = get_dataset_config_names( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py", line 347, in get_dataset_config_names dataset_module = dataset_module_factory( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1914, in dataset_module_factory raise e1 from None File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1880, in dataset_module_factory return HubDatasetModuleFactoryWithScript( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1504, in get_module local_imports = _download_additional_modules( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 354, in _download_additional_modules raise ImportError( ImportError: To be able to use universalner/universal_ner, you need to install the following dependency: conllu. Please install it using 'pip install conllu' for instance.
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
Dataset Card for Universal NER
Dataset Summary
Universal NER (UNER) is an open, community-driven initiative aimed at creating gold-standard benchmarks for Named Entity Recognition (NER) across multiple languages. The primary objective of UNER is to offer high-quality, cross-lingually consistent annotations, thereby standardizing and advancing multilingual NER research. UNER v1 includes 19 datasets with named entity annotations, uniformly structured across 13 diverse languages.
Supported Tasks and Leaderboards
token-classification
: The dataset can be used to train token classification models of the NER variety. Some pre-trained models released as part of the UNER v1 release can be found at https://huggingface.co./universalner
Languages
The dataset contains data in the following languages:
- Cebuano (
ceb
) - Danish (
da
) - German (
de
) - English (
en
) - Croatian (
hr
) - Portuguese (
pt
) - Russian (
ru
) - Slovak (
sk
) - Serbian (
sr
) - Swedish (
sv
) - Tagalog (
tl
) - Chinese (
zh
)
Dataset Structure
Data Instances
An example from the UNER_English-PUD
test set looks as follows
{
"idx": "n01016-0002",
"text": "Several analysts have suggested Huawei is best placed to benefit from Samsung's setback.",
"tokens": [
"Several", "analysts", "have", "suggested", "Huawei",
"is", "best", "placed", "to", "benefit",
"from", "Samsung", "'s", "setback", "."
],
"ner_tags": [
"O", "O", "O", "O", "B-ORG",
"O", "O", "O", "O", "O",
"O", "B-ORG", "O", "O", "O"
],
"annotator": "blvns"
}
Data Fields
idx
: the ID uniquely identifying the sentence (instance), if available.text
: the full text of the sentence (instance)tokens
: the text of the sentence (instance) split into tokens. Note that this split is inhereted from Universal Dependenciesner_tags
: the NER tags associated with each one of thetokens
annotator
: the annotator who provided thener_tags
for this particular instance
Data Splits
TBD
Dataset Creation
Curation Rationale
TBD
Source Data
Initial Data Collection and Normalization
We selected the Universal Dependency (UD) corpora as the default base texts for annotation due to their extensive language coverage, pre-existing data collection, cleaning, tokenization, and permissive licensing. This choice accelerates our process by providing a robust foundation. By adding another annotation layer to the already detailed UD annotations, we facilitate verification within our project and enable comprehensive multilingual research across the entire NLP pipeline. Given that UD annotations operate at the word level, we adopted the BIO annotation schema (specifically IOB2). In this schema, words forming the beginning (B) or inside (I) part of an entity (X ∈ {PER, LOC, ORG}) are annotated accordingly, while all other words receive an O tag. To maintain consistency, we preserve UD's original tokenization.
Although UD serves as the default data source for UNER, the project is not restricted to UD corpora, particularly for languages not currently represented in UD. The primary requirement for inclusion in the UNER corpus is adherence to the UNER tagging guidelines. Additionally, we are open to converting existing NER efforts on UD treebanks to align with UNER. In this initial release, we have included four datasets transferred from other manual annotation efforts on UD sources (for DA, HR, ARABIZI, and SR).
Who are the source language producers?
This information can be found on per-dataset basis for each of the source Universal Dependencies datasets.
Annotations
Annotation process
The data has been annotated by
Who are the annotators?
For the initial UNER annotation effort, we recruited volunteers from the multilingual NLP community via academic networks and social media. The annotators were coordinated through a Slack workspace, with all contributors working on a voluntary basis. We assume that annotators are either native speakers of the language they annotate or possess a high level of proficiency, although no formal language tests were conducted. The selection of the 13 dataset languages in the first UNER release was driven by the availability of annotators. As the project evolves, we anticipate the inclusion of additional languages and datasets as more annotators become available.
Personal and Sensitive Information
TBD
Considerations for Using the Data
Social Impact of Dataset
TBD
Discussion of Biases
TBD
Other Known Limitations
TBD
Additional Information
Dataset Curators
List the people involved in collecting the dataset and their affiliation(s). If funding information is known, include it here.
Licensing Information
The UNER v1 is released under the terms of the Creative Commons Attribution-ShareAlike 4.0 International license
Citation Information
If you use this dataset, please cite the corresponding paper:
@inproceedings{
mayhew2024universal,
title={Universal NER: A Gold-Standard Multilingual Named Entity Recognition Benchmark},
author={Stephen Mayhew and Terra Blevins and Shuheng Liu and Marek Šuppa and Hila Gonen and Joseph Marvin Imperial and Börje F. Karlsson and Peiqin Lin and Nikola Ljubešić and LJ Miranda and Barbara Plank and Arij Riab and Yuval Pinter}
booktitle={Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics (NAACL)},
year={2024},
url={https://aclanthology.org/2024.naacl-long.243/}
}
- Downloads last month
- 588