Datasets:

Modalities:
Text
Formats:
parquet
Libraries:
Datasets
Dask
License:
opus_paracrawl / README.md
albertvillanova's picture
Convert dataset to Parquet (#2)
8bb0bcf verified
metadata
annotations_creators:
  - found
language_creators:
  - found
language:
  - bg
  - ca
  - cs
  - da
  - de
  - el
  - en
  - es
  - et
  - eu
  - fi
  - fr
  - ga
  - gl
  - hr
  - hu
  - is
  - it
  - km
  - ko
  - lt
  - lv
  - mt
  - my
  - nb
  - ne
  - nl
  - nn
  - pl
  - pt
  - ro
  - ru
  - si
  - sk
  - sl
  - so
  - sv
  - sw
  - tl
  - uk
  - zh
license:
  - cc0-1.0
multilinguality:
  - multilingual
size_categories:
  - 100K<n<1M
  - 10K<n<100K
  - 1M<n<10M
source_datasets:
  - original
task_categories:
  - translation
task_ids: []
pretty_name: OpusParaCrawl
config_names:
  - de-pl
  - el-en
  - en-ha
  - en-ig
  - en-km
  - en-so
  - en-sw
  - en-tl
  - es-gl
  - fr-nl
dataset_info:
  - config_name: de-pl
    features:
      - name: id
        dtype: string
      - name: translation
        dtype:
          translation:
            languages:
              - de
              - pl
    splits:
      - name: train
        num_bytes: 298635927
        num_examples: 916643
    download_size: 183957290
    dataset_size: 298635927
  - config_name: el-en
    features:
      - name: id
        dtype: string
      - name: translation
        dtype:
          translation:
            languages:
              - el
              - en
    splits:
      - name: train
        num_bytes: 6760349369
        num_examples: 21402471
    download_size: 4108379167
    dataset_size: 6760349369
  - config_name: en-ha
    features:
      - name: id
        dtype: string
      - name: translation
        dtype:
          translation:
            languages:
              - en
              - ha
    splits:
      - name: train
        num_bytes: 4618460
        num_examples: 19694
    download_size: 1757433
    dataset_size: 4618460
  - config_name: en-ig
    features:
      - name: id
        dtype: string
      - name: translation
        dtype:
          translation:
            languages:
              - en
              - ig
    splits:
      - name: train
        num_bytes: 6709030
        num_examples: 28829
    download_size: 2691716
    dataset_size: 6709030
  - config_name: en-km
    features:
      - name: id
        dtype: string
      - name: translation
        dtype:
          translation:
            languages:
              - en
              - km
    splits:
      - name: train
        num_bytes: 31964409
        num_examples: 65115
    download_size: 16582595
    dataset_size: 31964409
  - config_name: en-so
    features:
      - name: id
        dtype: string
      - name: translation
        dtype:
          translation:
            languages:
              - en
              - so
    splits:
      - name: train
        num_bytes: 5790979
        num_examples: 14880
    download_size: 3718608
    dataset_size: 5790979
  - config_name: en-sw
    features:
      - name: id
        dtype: string
      - name: translation
        dtype:
          translation:
            languages:
              - en
              - sw
    splits:
      - name: train
        num_bytes: 44264274
        num_examples: 132520
    download_size: 30553316
    dataset_size: 44264274
  - config_name: en-tl
    features:
      - name: id
        dtype: string
      - name: translation
        dtype:
          translation:
            languages:
              - en
              - tl
    splits:
      - name: train
        num_bytes: 82502498
        num_examples: 248689
    download_size: 54686324
    dataset_size: 82502498
  - config_name: es-gl
    features:
      - name: id
        dtype: string
      - name: translation
        dtype:
          translation:
            languages:
              - es
              - gl
    splits:
      - name: train
        num_bytes: 582658645
        num_examples: 1879689
    download_size: 406732310
    dataset_size: 582658645
  - config_name: fr-nl
    features:
      - name: id
        dtype: string
      - name: translation
        dtype:
          translation:
            languages:
              - fr
              - nl
    splits:
      - name: train
        num_bytes: 862299992
        num_examples: 2687673
    download_size: 550812954
    dataset_size: 862299992
configs:
  - config_name: de-pl
    data_files:
      - split: train
        path: de-pl/train-*
  - config_name: el-en
    data_files:
      - split: train
        path: el-en/train-*
  - config_name: en-km
    data_files:
      - split: train
        path: en-km/train-*
  - config_name: en-so
    data_files:
      - split: train
        path: en-so/train-*
  - config_name: en-sw
    data_files:
      - split: train
        path: en-sw/train-*
  - config_name: en-tl
    data_files:
      - split: train
        path: en-tl/train-*
  - config_name: es-gl
    data_files:
      - split: train
        path: es-gl/train-*
  - config_name: fr-nl
    data_files:
      - split: train
        path: fr-nl/train-*

Dataset Card for OpusParaCrawl

Table of Contents

Dataset Description

Dataset Summary

Parallel corpora from Web Crawls collected in the ParaCrawl project.

Tha dataset contains:

  • 42 languages, 43 bitexts
  • total number of files: 59,996
  • total number of tokens: 56.11G
  • total number of sentence fragments: 3.13G

To load a language pair which isn't part of the config, all you need to do is specify the language code as pairs, e.g.

dataset = load_dataset("opus_paracrawl", lang1="en", lang2="so")

You can find the valid pairs in Homepage section of Dataset Description: http://opus.nlpl.eu/ParaCrawl.php

Supported Tasks and Leaderboards

[More Information Needed]

Languages

The languages in the dataset are:

  • bg
  • ca
  • cs
  • da
  • de
  • el
  • en
  • es
  • et
  • eu
  • fi
  • fr
  • ga
  • gl
  • hr
  • hu
  • is
  • it
  • km
  • ko
  • lt
  • lv
  • mt
  • my
  • nb
  • ne
  • nl
  • nn
  • pl
  • pt
  • ro
  • ru
  • si
  • sk
  • sl
  • so
  • sv
  • sw
  • tl
  • uk
  • zh

Dataset Structure

Data Instances

{
  'id': '0', 
  'translation': {
    "el": "Συνεχίστε ευθεία 300 μέτρα μέχρι να καταλήξουμε σε μια σωστή οδός (ul. Gagarina)? Περπατήστε περίπου 300 μέτρα μέχρι να φτάσετε το πρώτο ορθή οδός (ul Khotsa Namsaraeva)?",
    "en": "Go straight 300 meters until you come to a proper street (ul. Gagarina); Walk approximately 300 meters until you reach the first proper street (ul Khotsa Namsaraeva);"
  }
}

Data Fields

  • id (str): Unique identifier of the parallel sentence for the pair of languages.
  • translation (dict): Parallel sentences for the pair of languages.

Data Splits

The dataset contains a single train split.

Dataset Creation

Curation Rationale

[More Information Needed]

Source Data

[More Information Needed]

Initial Data Collection and Normalization

[More Information Needed]

Who are the source language producers?

[More Information Needed]

Annotations

[More Information Needed]

Annotation process

[More Information Needed]

Who are the annotators?

[More Information Needed]

Personal and Sensitive Information

[More Information Needed]

Considerations for Using the Data

Social Impact of Dataset

[More Information Needed]

Discussion of Biases

[More Information Needed]

Other Known Limitations

[More Information Needed]

Additional Information

Dataset Curators

[More Information Needed]

Licensing Information

  • Creative commons CC0 (no rights reserved)

Citation Information

@inproceedings{banon-etal-2020-paracrawl,
    title = "{P}ara{C}rawl: Web-Scale Acquisition of Parallel Corpora",
    author = "Ba{\~n}{\'o}n, Marta  and
      Chen, Pinzhen  and
      Haddow, Barry  and
      Heafield, Kenneth  and
      Hoang, Hieu  and
      Espl{\`a}-Gomis, Miquel  and
      Forcada, Mikel L.  and
      Kamran, Amir  and
      Kirefu, Faheem  and
      Koehn, Philipp  and
      Ortiz Rojas, Sergio  and
      Pla Sempere, Leopoldo  and
      Ram{\'\i}rez-S{\'a}nchez, Gema  and
      Sarr{\'\i}as, Elsa  and
      Strelec, Marek  and
      Thompson, Brian  and
      Waites, William  and
      Wiggins, Dion  and
      Zaragoza, Jaume",
    booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics",
    month = jul,
    year = "2020",
    address = "Online",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2020.acl-main.417",
    doi = "10.18653/v1/2020.acl-main.417",
    pages = "4555--4567",
}
@InProceedings{TIEDEMANN12.463,
  author = {Jörg Tiedemann},
  title = {Parallel Data, Tools and Interfaces in OPUS},
  booktitle = {Proceedings of the Eight International Conference on Language Resources and Evaluation (LREC'12)},
  year = {2012},
  month = {may},
  date = {23-25},
  address = {Istanbul, Turkey},
  editor = {Nicoletta Calzolari (Conference Chair) and Khalid Choukri and Thierry Declerck and Mehmet Uğur Doğan and Bente Maegaard and Joseph Mariani and Asuncion Moreno and Jan Odijk and Stelios Piperidis},
  publisher = {European Language Resources Association (ELRA)},
  isbn = {978-2-9517408-7-7},
  language = {english}
}

Contributions

Thanks to @rkc007 for adding this dataset.