Datasets:
GEM
/

ArXiv:
License:
Ronald Cardenas Acosta commited on
Commit
031583f
1 Parent(s): 5e8924f

new language pairs zh-* and *-zh

Browse files
Files changed (6) hide show
  1. .gitattributes +2 -0
  2. README.md +1 -1
  3. test/zh.jsonl +2 -2
  4. train/zh.jsonl +3 -0
  5. valid/zh.jsonl +3 -0
  6. xwikis.py +6 -37
.gitattributes CHANGED
@@ -93,3 +93,5 @@ valid/cs-zh.jsonl filter=lfs diff=lfs merge=lfs -text
93
  valid/de-zh.jsonl filter=lfs diff=lfs merge=lfs -text
94
  valid/en-zh.jsonl filter=lfs diff=lfs merge=lfs -text
95
  valid/fr-zh.jsonl filter=lfs diff=lfs merge=lfs -text
 
 
 
93
  valid/de-zh.jsonl filter=lfs diff=lfs merge=lfs -text
94
  valid/en-zh.jsonl filter=lfs diff=lfs merge=lfs -text
95
  valid/fr-zh.jsonl filter=lfs diff=lfs merge=lfs -text
96
+ train/zh.jsonl filter=lfs diff=lfs merge=lfs -text
97
+ valid/zh.jsonl filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -120,7 +120,7 @@ yes
120
  <!-- quick -->
121
  <!-- info: What languages/dialects are covered in the dataset? -->
122
  <!-- scope: telescope -->
123
- `German`, `English`, `French`, `Czech`
124
 
125
  #### License
126
 
 
120
  <!-- quick -->
121
  <!-- info: What languages/dialects are covered in the dataset? -->
122
  <!-- scope: telescope -->
123
+ `German`, `English`, `French`, `Czech`, `Chinese`
124
 
125
  #### License
126
 
test/zh.jsonl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e07db118beaafcbacae3751bcb15555e8760af579e62b1fc413398c2bbf2e91
3
- size 37178864
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef69fcfc6ed855373055a59303ea892f0d46f115688aa883e8a3d492d31898f2
3
+ size 37214675
train/zh.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:344beffc3daf2abd9535956fd687ab43ea893bfa95191605f0f98ce009d57de4
3
+ size 1065902922
valid/zh.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2a784e4d7c1deabe877db71a3978b8194f80e84d616d1e1a544745752a2fa60
3
+ size 96364048
xwikis.py CHANGED
@@ -18,7 +18,7 @@
18
  import csv
19
  import json
20
  import os
21
-
22
  import datasets
23
 
24
 
@@ -51,47 +51,16 @@ VERSION = datasets.Version("0.1.0")
51
  # TODO: Add link to the official dataset URLs here
52
  # The HuggingFace dataset library don't host the datasets but only point to the original files
53
  # This can be an arbitrary nested dict/list of URLs (see below in `_split_generators` method)
 
54
 
55
  LPAIRS = [
56
- f"{x}-{y}" for x in ["en","fr","cs","de"] for y in ["en","fr","cs","de"] if x!=y
57
  ]
58
 
59
  _URLs = {
60
- "train": [ f"./train/{xy}.jsonl" for xy in LPAIRS],
61
- # "./train/en-fr.jsonl",
62
- # "./train/fr-en.jsonl",
63
- # "./train/en-de.jsonl",
64
- # "./train/de-en.jsonl",
65
- # "./train/en-cs.jsonl",
66
- # "./train/cs-en.jsonl",
67
- # "./train/fr-de.jsonl",
68
- # "./train/de-fr.jsonl",
69
- # "./train/fr-cs.jsonl",
70
- # "./train/cs-fr.jsonl",
71
- # "./train/de-cs.jsonl",
72
- # "./train/cs-de.jsonl",
73
- # ],
74
- "validation": [ f"./valid/{xy}.jsonl" for xy in LPAIRS],
75
- # "validation": [
76
- # "./valid/en-fr.jsonl",
77
- # "./valid/fr-en.jsonl",
78
- # "./valid/en-de.jsonl",
79
- # "./valid/de-en.jsonl",
80
- # "./valid/en-cs.jsonl",
81
- # "./valid/cs-en.jsonl",
82
- # "./valid/fr-de.jsonl",
83
- # "./valid/de-fr.jsonl",
84
- # "./valid/fr-cs.jsonl",
85
- # "./valid/cs-fr.jsonl",
86
- # "./valid/de-cs.jsonl",
87
- # "./valid/cs-de.jsonl",
88
- # ],
89
- "test": [ f"./test/{xy}.jsonl" for xy in LPAIRS if "-en" in xy],
90
- # "test": [
91
- # "./test/fr-en.jsonl",
92
- # "./test/de-en.jsonl",
93
- # "./test/cs-en.jsonl",
94
- # ],
95
  }
96
 
97
 
 
18
  import csv
19
  import json
20
  import os
21
+ import itertools
22
  import datasets
23
 
24
 
 
51
  # TODO: Add link to the official dataset URLs here
52
  # The HuggingFace dataset library don't host the datasets but only point to the original files
53
  # This can be an arbitrary nested dict/list of URLs (see below in `_split_generators` method)
54
+ LANGS = ["en","fr","cs","de","zh"]
55
 
56
  LPAIRS = [
57
+ f"{x}-{y}" for x,y in itertools.product(LANGS,LANGS) if x!=y
58
  ]
59
 
60
  _URLs = {
61
+ "train": [ f"./train/{xy}.jsonl" for xy in LPAIRS] + [ f"./train/{x}.jsonl" for x in LANGS],
62
+ "validation": [ f"./valid/{xy}.jsonl" for xy in LPAIRS] + [ f"./valid/{x}.jsonl" for x in LANGS],
63
+ "test": [ f"./test/{xy}.jsonl" for xy in LPAIRS if "-en" in xy] + [ f"./test/{x}.jsonl" for x in LANGS],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
64
  }
65
 
66