Convert dataset to Parquet (#4)
Browse files- Convert dataset to Parquet (bf446c28e394d34942439ece305066d8e000cc4b)
- Add emotion data files (2d948f1722cd37b0958acedbbb4c50df5d3783d0)
- Add hate data files (02a5ceb555b982fa24beb13722977f1b64013039)
- Add irony data files (d0b81754adc6f44a439a2866dada5a1f0f1b55f6)
- Add offensive data files (9b98abef0b187e34801efc69f20641e9b60a0961)
- Add sentiment data files (6fe62d6334fb7f8197508d057f6351333dbe7646)
- Add stance_abortion data files (55864a3201f347e9edfac045e7f772eb2c0ee801)
- Add stance_atheism data files (449c8d551b86ba45c3c37b511f1f8d7ea91ec3ec)
- Add stance_climate data files (ff01bee78da3a8ec1c5ab3bf7d9d4c8f5121f24f)
- Add stance_feminist data files (5ea3f6097320422a56ac6677a992bb34fd58420f)
- Add stance_hillary data files (c460af45e62a2692f21d9755cfe8b3b9de801b4a)
- Delete loading script (68be50101769462a8a0b5fa46572a84b7d78d72e)
- Delete legacy dataset_infos.json (684f8540f99a7aad647599b97a3219ddc2801c0a)
- README.md +156 -67
- dataset_infos.json +0 -1
- emoji/test-00000-of-00001.parquet +3 -0
- emoji/train-00000-of-00001.parquet +3 -0
- emoji/validation-00000-of-00001.parquet +3 -0
- emotion/test-00000-of-00001.parquet +3 -0
- emotion/train-00000-of-00001.parquet +3 -0
- emotion/validation-00000-of-00001.parquet +3 -0
- hate/test-00000-of-00001.parquet +3 -0
- hate/train-00000-of-00001.parquet +3 -0
- hate/validation-00000-of-00001.parquet +3 -0
- irony/test-00000-of-00001.parquet +3 -0
- irony/train-00000-of-00001.parquet +3 -0
- irony/validation-00000-of-00001.parquet +3 -0
- offensive/test-00000-of-00001.parquet +3 -0
- offensive/train-00000-of-00001.parquet +3 -0
- offensive/validation-00000-of-00001.parquet +3 -0
- sentiment/test-00000-of-00001.parquet +3 -0
- sentiment/train-00000-of-00001.parquet +3 -0
- sentiment/validation-00000-of-00001.parquet +3 -0
- stance_abortion/test-00000-of-00001.parquet +3 -0
- stance_abortion/train-00000-of-00001.parquet +3 -0
- stance_abortion/validation-00000-of-00001.parquet +3 -0
- stance_atheism/test-00000-of-00001.parquet +3 -0
- stance_atheism/train-00000-of-00001.parquet +3 -0
- stance_atheism/validation-00000-of-00001.parquet +3 -0
- stance_climate/test-00000-of-00001.parquet +3 -0
- stance_climate/train-00000-of-00001.parquet +3 -0
- stance_climate/validation-00000-of-00001.parquet +3 -0
- stance_feminist/test-00000-of-00001.parquet +3 -0
- stance_feminist/train-00000-of-00001.parquet +3 -0
- stance_feminist/validation-00000-of-00001.parquet +3 -0
- stance_hillary/test-00000-of-00001.parquet +3 -0
- stance_hillary/train-00000-of-00001.parquet +3 -0
- stance_hillary/validation-00000-of-00001.parquet +3 -0
- tweet_eval.py +0 -249
@@ -24,6 +24,18 @@ task_ids:
|
|
24 |
- sentiment-classification
|
25 |
paperswithcode_id: tweeteval
|
26 |
pretty_name: TweetEval
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
27 |
dataset_info:
|
28 |
- config_name: emoji
|
29 |
features:
|
@@ -55,16 +67,16 @@ dataset_info:
|
|
55 |
'19': π
|
56 |
splits:
|
57 |
- name: train
|
58 |
-
num_bytes:
|
59 |
num_examples: 45000
|
60 |
- name: test
|
61 |
-
num_bytes:
|
62 |
num_examples: 50000
|
63 |
- name: validation
|
64 |
-
num_bytes:
|
65 |
num_examples: 5000
|
66 |
-
download_size:
|
67 |
-
dataset_size:
|
68 |
- config_name: emotion
|
69 |
features:
|
70 |
- name: text
|
@@ -79,16 +91,16 @@ dataset_info:
|
|
79 |
'3': sadness
|
80 |
splits:
|
81 |
- name: train
|
82 |
-
num_bytes:
|
83 |
num_examples: 3257
|
84 |
- name: test
|
85 |
-
num_bytes:
|
86 |
num_examples: 1421
|
87 |
- name: validation
|
88 |
-
num_bytes:
|
89 |
num_examples: 374
|
90 |
-
download_size:
|
91 |
-
dataset_size:
|
92 |
- config_name: hate
|
93 |
features:
|
94 |
- name: text
|
@@ -101,16 +113,16 @@ dataset_info:
|
|
101 |
'1': hate
|
102 |
splits:
|
103 |
- name: train
|
104 |
-
num_bytes:
|
105 |
num_examples: 9000
|
106 |
- name: test
|
107 |
-
num_bytes:
|
108 |
num_examples: 2970
|
109 |
- name: validation
|
110 |
-
num_bytes:
|
111 |
num_examples: 1000
|
112 |
-
download_size:
|
113 |
-
dataset_size:
|
114 |
- config_name: irony
|
115 |
features:
|
116 |
- name: text
|
@@ -123,16 +135,16 @@ dataset_info:
|
|
123 |
'1': irony
|
124 |
splits:
|
125 |
- name: train
|
126 |
-
num_bytes:
|
127 |
num_examples: 2862
|
128 |
- name: test
|
129 |
-
num_bytes:
|
130 |
num_examples: 784
|
131 |
- name: validation
|
132 |
-
num_bytes:
|
133 |
num_examples: 955
|
134 |
-
download_size:
|
135 |
-
dataset_size:
|
136 |
- config_name: offensive
|
137 |
features:
|
138 |
- name: text
|
@@ -145,16 +157,16 @@ dataset_info:
|
|
145 |
'1': offensive
|
146 |
splits:
|
147 |
- name: train
|
148 |
-
num_bytes:
|
149 |
num_examples: 11916
|
150 |
- name: test
|
151 |
-
num_bytes:
|
152 |
num_examples: 860
|
153 |
- name: validation
|
154 |
-
num_bytes:
|
155 |
num_examples: 1324
|
156 |
-
download_size:
|
157 |
-
dataset_size:
|
158 |
- config_name: sentiment
|
159 |
features:
|
160 |
- name: text
|
@@ -168,16 +180,16 @@ dataset_info:
|
|
168 |
'2': positive
|
169 |
splits:
|
170 |
- name: train
|
171 |
-
num_bytes:
|
172 |
num_examples: 45615
|
173 |
- name: test
|
174 |
-
num_bytes:
|
175 |
num_examples: 12284
|
176 |
- name: validation
|
177 |
-
num_bytes:
|
178 |
num_examples: 2000
|
179 |
-
download_size:
|
180 |
-
dataset_size:
|
181 |
- config_name: stance_abortion
|
182 |
features:
|
183 |
- name: text
|
@@ -191,16 +203,16 @@ dataset_info:
|
|
191 |
'2': favor
|
192 |
splits:
|
193 |
- name: train
|
194 |
-
num_bytes:
|
195 |
num_examples: 587
|
196 |
- name: test
|
197 |
-
num_bytes:
|
198 |
num_examples: 280
|
199 |
- name: validation
|
200 |
-
num_bytes:
|
201 |
num_examples: 66
|
202 |
-
download_size:
|
203 |
-
dataset_size:
|
204 |
- config_name: stance_atheism
|
205 |
features:
|
206 |
- name: text
|
@@ -214,16 +226,16 @@ dataset_info:
|
|
214 |
'2': favor
|
215 |
splits:
|
216 |
- name: train
|
217 |
-
num_bytes:
|
218 |
num_examples: 461
|
219 |
- name: test
|
220 |
-
num_bytes:
|
221 |
num_examples: 220
|
222 |
- name: validation
|
223 |
-
num_bytes:
|
224 |
num_examples: 52
|
225 |
-
download_size:
|
226 |
-
dataset_size:
|
227 |
- config_name: stance_climate
|
228 |
features:
|
229 |
- name: text
|
@@ -237,16 +249,16 @@ dataset_info:
|
|
237 |
'2': favor
|
238 |
splits:
|
239 |
- name: train
|
240 |
-
num_bytes:
|
241 |
num_examples: 355
|
242 |
- name: test
|
243 |
-
num_bytes:
|
244 |
num_examples: 169
|
245 |
- name: validation
|
246 |
-
num_bytes:
|
247 |
num_examples: 40
|
248 |
-
download_size:
|
249 |
-
dataset_size:
|
250 |
- config_name: stance_feminist
|
251 |
features:
|
252 |
- name: text
|
@@ -260,16 +272,16 @@ dataset_info:
|
|
260 |
'2': favor
|
261 |
splits:
|
262 |
- name: train
|
263 |
-
num_bytes:
|
264 |
num_examples: 597
|
265 |
- name: test
|
266 |
-
num_bytes:
|
267 |
num_examples: 285
|
268 |
- name: validation
|
269 |
-
num_bytes:
|
270 |
num_examples: 67
|
271 |
-
download_size:
|
272 |
-
dataset_size:
|
273 |
- config_name: stance_hillary
|
274 |
features:
|
275 |
- name: text
|
@@ -283,16 +295,105 @@ dataset_info:
|
|
283 |
'2': favor
|
284 |
splits:
|
285 |
- name: train
|
286 |
-
num_bytes:
|
287 |
num_examples: 620
|
288 |
- name: test
|
289 |
-
num_bytes:
|
290 |
num_examples: 295
|
291 |
- name: validation
|
292 |
-
num_bytes:
|
293 |
num_examples: 69
|
294 |
-
download_size:
|
295 |
-
dataset_size:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
296 |
train-eval-index:
|
297 |
- config: emotion
|
298 |
task: text-classification
|
@@ -510,18 +611,6 @@ train-eval-index:
|
|
510 |
name: Recall weighted
|
511 |
args:
|
512 |
average: weighted
|
513 |
-
config_names:
|
514 |
-
- emoji
|
515 |
-
- emotion
|
516 |
-
- hate
|
517 |
-
- irony
|
518 |
-
- offensive
|
519 |
-
- sentiment
|
520 |
-
- stance_abortion
|
521 |
-
- stance_atheism
|
522 |
-
- stance_climate
|
523 |
-
- stance_feminist
|
524 |
-
- stance_hillary
|
525 |
---
|
526 |
|
527 |
# Dataset Card for tweet_eval
|
|
|
24 |
- sentiment-classification
|
25 |
paperswithcode_id: tweeteval
|
26 |
pretty_name: TweetEval
|
27 |
+
config_names:
|
28 |
+
- emoji
|
29 |
+
- emotion
|
30 |
+
- hate
|
31 |
+
- irony
|
32 |
+
- offensive
|
33 |
+
- sentiment
|
34 |
+
- stance_abortion
|
35 |
+
- stance_atheism
|
36 |
+
- stance_climate
|
37 |
+
- stance_feminist
|
38 |
+
- stance_hillary
|
39 |
dataset_info:
|
40 |
- config_name: emoji
|
41 |
features:
|
|
|
67 |
'19': π
|
68 |
splits:
|
69 |
- name: train
|
70 |
+
num_bytes: 3803167
|
71 |
num_examples: 45000
|
72 |
- name: test
|
73 |
+
num_bytes: 4255901
|
74 |
num_examples: 50000
|
75 |
- name: validation
|
76 |
+
num_bytes: 396079
|
77 |
num_examples: 5000
|
78 |
+
download_size: 5939308
|
79 |
+
dataset_size: 8455147
|
80 |
- config_name: emotion
|
81 |
features:
|
82 |
- name: text
|
|
|
91 |
'3': sadness
|
92 |
splits:
|
93 |
- name: train
|
94 |
+
num_bytes: 338871
|
95 |
num_examples: 3257
|
96 |
- name: test
|
97 |
+
num_bytes: 146645
|
98 |
num_examples: 1421
|
99 |
- name: validation
|
100 |
+
num_bytes: 38273
|
101 |
num_examples: 374
|
102 |
+
download_size: 367016
|
103 |
+
dataset_size: 523789
|
104 |
- config_name: hate
|
105 |
features:
|
106 |
- name: text
|
|
|
113 |
'1': hate
|
114 |
splits:
|
115 |
- name: train
|
116 |
+
num_bytes: 1223650
|
117 |
num_examples: 9000
|
118 |
- name: test
|
119 |
+
num_bytes: 428934
|
120 |
num_examples: 2970
|
121 |
- name: validation
|
122 |
+
num_bytes: 154144
|
123 |
num_examples: 1000
|
124 |
+
download_size: 1196346
|
125 |
+
dataset_size: 1806728
|
126 |
- config_name: irony
|
127 |
features:
|
128 |
- name: text
|
|
|
135 |
'1': irony
|
136 |
splits:
|
137 |
- name: train
|
138 |
+
num_bytes: 259187
|
139 |
num_examples: 2862
|
140 |
- name: test
|
141 |
+
num_bytes: 75897
|
142 |
num_examples: 784
|
143 |
- name: validation
|
144 |
+
num_bytes: 86017
|
145 |
num_examples: 955
|
146 |
+
download_size: 297647
|
147 |
+
dataset_size: 421101
|
148 |
- config_name: offensive
|
149 |
features:
|
150 |
- name: text
|
|
|
157 |
'1': offensive
|
158 |
splits:
|
159 |
- name: train
|
160 |
+
num_bytes: 1648061
|
161 |
num_examples: 11916
|
162 |
- name: test
|
163 |
+
num_bytes: 135473
|
164 |
num_examples: 860
|
165 |
- name: validation
|
166 |
+
num_bytes: 192417
|
167 |
num_examples: 1324
|
168 |
+
download_size: 1234528
|
169 |
+
dataset_size: 1975951
|
170 |
- config_name: sentiment
|
171 |
features:
|
172 |
- name: text
|
|
|
180 |
'2': positive
|
181 |
splits:
|
182 |
- name: train
|
183 |
+
num_bytes: 5425122
|
184 |
num_examples: 45615
|
185 |
- name: test
|
186 |
+
num_bytes: 1279540
|
187 |
num_examples: 12284
|
188 |
- name: validation
|
189 |
+
num_bytes: 239084
|
190 |
num_examples: 2000
|
191 |
+
download_size: 4849675
|
192 |
+
dataset_size: 6943746
|
193 |
- config_name: stance_abortion
|
194 |
features:
|
195 |
- name: text
|
|
|
203 |
'2': favor
|
204 |
splits:
|
205 |
- name: train
|
206 |
+
num_bytes: 68694
|
207 |
num_examples: 587
|
208 |
- name: test
|
209 |
+
num_bytes: 33171
|
210 |
num_examples: 280
|
211 |
- name: validation
|
212 |
+
num_bytes: 7657
|
213 |
num_examples: 66
|
214 |
+
download_size: 73517
|
215 |
+
dataset_size: 109522
|
216 |
- config_name: stance_atheism
|
217 |
features:
|
218 |
- name: text
|
|
|
226 |
'2': favor
|
227 |
splits:
|
228 |
- name: train
|
229 |
+
num_bytes: 54775
|
230 |
num_examples: 461
|
231 |
- name: test
|
232 |
+
num_bytes: 25716
|
233 |
num_examples: 220
|
234 |
- name: validation
|
235 |
+
num_bytes: 6320
|
236 |
num_examples: 52
|
237 |
+
download_size: 62265
|
238 |
+
dataset_size: 86811
|
239 |
- config_name: stance_climate
|
240 |
features:
|
241 |
- name: text
|
|
|
249 |
'2': favor
|
250 |
splits:
|
251 |
- name: train
|
252 |
+
num_bytes: 40249
|
253 |
num_examples: 355
|
254 |
- name: test
|
255 |
+
num_bytes: 19925
|
256 |
num_examples: 169
|
257 |
- name: validation
|
258 |
+
num_bytes: 4801
|
259 |
num_examples: 40
|
260 |
+
download_size: 48493
|
261 |
+
dataset_size: 64975
|
262 |
- config_name: stance_feminist
|
263 |
features:
|
264 |
- name: text
|
|
|
272 |
'2': favor
|
273 |
splits:
|
274 |
- name: train
|
275 |
+
num_bytes: 70509
|
276 |
num_examples: 597
|
277 |
- name: test
|
278 |
+
num_bytes: 33305
|
279 |
num_examples: 285
|
280 |
- name: validation
|
281 |
+
num_bytes: 8035
|
282 |
num_examples: 67
|
283 |
+
download_size: 76345
|
284 |
+
dataset_size: 111849
|
285 |
- config_name: stance_hillary
|
286 |
features:
|
287 |
- name: text
|
|
|
295 |
'2': favor
|
296 |
splits:
|
297 |
- name: train
|
298 |
+
num_bytes: 69596
|
299 |
num_examples: 620
|
300 |
- name: test
|
301 |
+
num_bytes: 34487
|
302 |
num_examples: 295
|
303 |
- name: validation
|
304 |
+
num_bytes: 7532
|
305 |
num_examples: 69
|
306 |
+
download_size: 74057
|
307 |
+
dataset_size: 111615
|
308 |
+
configs:
|
309 |
+
- config_name: emoji
|
310 |
+
data_files:
|
311 |
+
- split: train
|
312 |
+
path: emoji/train-*
|
313 |
+
- split: test
|
314 |
+
path: emoji/test-*
|
315 |
+
- split: validation
|
316 |
+
path: emoji/validation-*
|
317 |
+
- config_name: emotion
|
318 |
+
data_files:
|
319 |
+
- split: train
|
320 |
+
path: emotion/train-*
|
321 |
+
- split: test
|
322 |
+
path: emotion/test-*
|
323 |
+
- split: validation
|
324 |
+
path: emotion/validation-*
|
325 |
+
- config_name: hate
|
326 |
+
data_files:
|
327 |
+
- split: train
|
328 |
+
path: hate/train-*
|
329 |
+
- split: test
|
330 |
+
path: hate/test-*
|
331 |
+
- split: validation
|
332 |
+
path: hate/validation-*
|
333 |
+
- config_name: irony
|
334 |
+
data_files:
|
335 |
+
- split: train
|
336 |
+
path: irony/train-*
|
337 |
+
- split: test
|
338 |
+
path: irony/test-*
|
339 |
+
- split: validation
|
340 |
+
path: irony/validation-*
|
341 |
+
- config_name: offensive
|
342 |
+
data_files:
|
343 |
+
- split: train
|
344 |
+
path: offensive/train-*
|
345 |
+
- split: test
|
346 |
+
path: offensive/test-*
|
347 |
+
- split: validation
|
348 |
+
path: offensive/validation-*
|
349 |
+
- config_name: sentiment
|
350 |
+
data_files:
|
351 |
+
- split: train
|
352 |
+
path: sentiment/train-*
|
353 |
+
- split: test
|
354 |
+
path: sentiment/test-*
|
355 |
+
- split: validation
|
356 |
+
path: sentiment/validation-*
|
357 |
+
- config_name: stance_abortion
|
358 |
+
data_files:
|
359 |
+
- split: train
|
360 |
+
path: stance_abortion/train-*
|
361 |
+
- split: test
|
362 |
+
path: stance_abortion/test-*
|
363 |
+
- split: validation
|
364 |
+
path: stance_abortion/validation-*
|
365 |
+
- config_name: stance_atheism
|
366 |
+
data_files:
|
367 |
+
- split: train
|
368 |
+
path: stance_atheism/train-*
|
369 |
+
- split: test
|
370 |
+
path: stance_atheism/test-*
|
371 |
+
- split: validation
|
372 |
+
path: stance_atheism/validation-*
|
373 |
+
- config_name: stance_climate
|
374 |
+
data_files:
|
375 |
+
- split: train
|
376 |
+
path: stance_climate/train-*
|
377 |
+
- split: test
|
378 |
+
path: stance_climate/test-*
|
379 |
+
- split: validation
|
380 |
+
path: stance_climate/validation-*
|
381 |
+
- config_name: stance_feminist
|
382 |
+
data_files:
|
383 |
+
- split: train
|
384 |
+
path: stance_feminist/train-*
|
385 |
+
- split: test
|
386 |
+
path: stance_feminist/test-*
|
387 |
+
- split: validation
|
388 |
+
path: stance_feminist/validation-*
|
389 |
+
- config_name: stance_hillary
|
390 |
+
data_files:
|
391 |
+
- split: train
|
392 |
+
path: stance_hillary/train-*
|
393 |
+
- split: test
|
394 |
+
path: stance_hillary/test-*
|
395 |
+
- split: validation
|
396 |
+
path: stance_hillary/validation-*
|
397 |
train-eval-index:
|
398 |
- config: emotion
|
399 |
task: text-classification
|
|
|
611 |
name: Recall weighted
|
612 |
args:
|
613 |
average: weighted
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
614 |
---
|
615 |
|
616 |
# Dataset Card for tweet_eval
|
@@ -1 +0,0 @@
|
|
1 |
-
{"emoji": {"description": "TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.\n", "citation": "@inproceedings{barbieri2020tweeteval,\n title={{TweetEval:Unified Benchmark and Comparative Evaluation for Tweet Classification}},\n author={Barbieri, Francesco and Camacho-Collados, Jose and Espinosa-Anke, Luis and Neves, Leonardo},\n booktitle={Proceedings of Findings of EMNLP},\n year={2020}\n}\n", "homepage": "https://github.com/cardiffnlp/tweeteval", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 20, "names": ["\u2764", "\ud83d\ude0d", "\ud83d\ude02", "\ud83d\udc95", "\ud83d\udd25", "\ud83d\ude0a", "\ud83d\ude0e", "\u2728", "\ud83d\udc99", "\ud83d\ude18", "\ud83d\udcf7", "\ud83c\uddfa\ud83c\uddf8", "\u2600", "\ud83d\udc9c", "\ud83d\ude09", "\ud83d\udcaf", "\ud83d\ude01", "\ud83c\udf84", "\ud83d\udcf8", "\ud83d\ude1c"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "builder_name": " tweet_eval", "config_name": "emoji", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 3803187, "num_examples": 45000, "dataset_name": " tweet_eval"}, "test": {"name": "test", "num_bytes": 4255921, "num_examples": 50000, "dataset_name": " tweet_eval"}, "validation": {"name": "validation", "num_bytes": 396083, "num_examples": 5000, "dataset_name": " tweet_eval"}}, "download_checksums": {"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/emoji/train_text.txt": {"num_bytes": 3353167, "checksum": "eacb6b0ee1fe2803d72a009c2e731fe07659f604318a979951d2f07c23c564a1"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/emoji/train_labels.txt": {"num_bytes": 102760, "checksum": "daee7da826683dbfa50ad3a29c60bc527e498f06c70eabee3745a99cc37ab3a5"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/emoji/test_text.txt": {"num_bytes": 3705901, "checksum": "e4de11de1597842c431dd67868e83322f5a432564dfd8558889ed8ac6a1a5e09"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/emoji/test_labels.txt": {"num_bytes": 114435, "checksum": "c1662b84788f36674ab8f0106f3e2e7d3e258ddf4959086ac7cc75b1e68dd1f6"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/emoji/val_text.txt": {"num_bytes": 341079, "checksum": "3bc3742d6af404cea792671878684d110f3bc02fd79a2e34643789a521d81a26"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/emoji/val_labels.txt": {"num_bytes": 11379, "checksum": "21ba456f688668d049ff0fb1fa04469ee684cf4e2467c71d2c3fe5ca2ba1bd1a"}}, "download_size": 7628721, "post_processing_size": null, "dataset_size": 8455191, "size_in_bytes": 16083912}, "emotion": {"description": "TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.\n", "citation": "@inproceedings{barbieri2020tweeteval,\n title={{TweetEval:Unified Benchmark and Comparative Evaluation for Tweet Classification}},\n author={Barbieri, Francesco and Camacho-Collados, Jose and Espinosa-Anke, Luis and Neves, Leonardo},\n booktitle={Proceedings of Findings of EMNLP},\n year={2020}\n}\n", "homepage": "https://github.com/cardiffnlp/tweeteval", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 4, "names": ["anger", "joy", "optimism", "sadness"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "builder_name": " tweet_eval", "config_name": "emotion", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 338875, "num_examples": 3257, "dataset_name": " tweet_eval"}, "test": {"name": "test", "num_bytes": 146649, "num_examples": 1421, "dataset_name": " tweet_eval"}, "validation": {"name": "validation", "num_bytes": 38277, "num_examples": 374, "dataset_name": " tweet_eval"}}, "download_checksums": {"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/emotion/train_text.txt": {"num_bytes": 306630, "checksum": "2c62f67aeb3eac1aea0e5a9c3d0f4bc337992581f3f858061786a1fb4d79d95e"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/emotion/train_labels.txt": {"num_bytes": 6514, "checksum": "987e767d8679e18abdf7de37a6d2bcd0a40a296ddd704e8d515cf0e3033c8d9c"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/emotion/test_text.txt": {"num_bytes": 132523, "checksum": "7e1070f5d3e3fcece5bc73680bff9981e90d8f7b2f1009bfe7a01d059d1c6091"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/emotion/test_labels.txt": {"num_bytes": 2842, "checksum": "245072348c711961785be6d395997f97cf7fcda3effeae7805664171dc75f913"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/emotion/val_text.txt": {"num_bytes": 34556, "checksum": "e2e30c86b8cbb97944d6543aedc06eace3bb275cb2f381aba787b838b4f23ca5"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/emotion/val_labels.txt": {"num_bytes": 748, "checksum": "313730630160b7e0a6b4235b800c76683f4aeeb72d094eb69646630cd5cfe338"}}, "download_size": 483813, "post_processing_size": null, "dataset_size": 523801, "size_in_bytes": 1007614}, "hate": {"description": "TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.\n", "citation": "@inproceedings{barbieri2020tweeteval,\n title={{TweetEval:Unified Benchmark and Comparative Evaluation for Tweet Classification}},\n author={Barbieri, Francesco and Camacho-Collados, Jose and Espinosa-Anke, Luis and Neves, Leonardo},\n booktitle={Proceedings of Findings of EMNLP},\n year={2020}\n}\n", "homepage": "https://github.com/cardiffnlp/tweeteval", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 2, "names": ["non-hate", "hate"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "builder_name": " tweet_eval", "config_name": "hate", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 1223654, "num_examples": 9000, "dataset_name": " tweet_eval"}, "test": {"name": "test", "num_bytes": 428938, "num_examples": 2970, "dataset_name": " tweet_eval"}, "validation": {"name": "validation", "num_bytes": 154148, "num_examples": 1000, "dataset_name": " tweet_eval"}}, "download_checksums": {"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/hate/train_text.txt": {"num_bytes": 1133852, "checksum": "6572bb3a42143128a5dfa99af8debeb0668e637c34b2d1e3140dac47316fe2c2"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/hate/train_labels.txt": {"num_bytes": 18000, "checksum": "4e8fde025a453a25c94632794254131dedeac4e57228ad64157c41571cc88f71"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/hate/test_text.txt": {"num_bytes": 399242, "checksum": "bc4762876a8dd8baa55c3cd7b03108e3231a5d691e80b8b1ef97c5be31b9da9a"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/hate/test_labels.txt": {"num_bytes": 5940, "checksum": "c14adca6b3627616a835c5ccea8a1cceb0235cd79417257f093eb0e16a69c62f"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/hate/val_text.txt": {"num_bytes": 144174, "checksum": "1ff78b1ed4c5ce43284b9eba32eb7d60c6d45d0d1b3b4d6df456ae01640764f1"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/hate/val_labels.txt": {"num_bytes": 2000, "checksum": "5092badf1a0e70036ea6264bcd0b78afc07d0f4a512fa6af34c2c4973600656b"}}, "download_size": 1703208, "post_processing_size": null, "dataset_size": 1806740, "size_in_bytes": 3509948}, "irony": {"description": "TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.\n", "citation": "@inproceedings{barbieri2020tweeteval,\n title={{TweetEval:Unified Benchmark and Comparative Evaluation for Tweet Classification}},\n author={Barbieri, Francesco and Camacho-Collados, Jose and Espinosa-Anke, Luis and Neves, Leonardo},\n booktitle={Proceedings of Findings of EMNLP},\n year={2020}\n}\n", "homepage": "https://github.com/cardiffnlp/tweeteval", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 2, "names": ["non_irony", "irony"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "builder_name": " tweet_eval", "config_name": "irony", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 259191, "num_examples": 2862, "dataset_name": " tweet_eval"}, "test": {"name": "test", "num_bytes": 75901, "num_examples": 784, "dataset_name": " tweet_eval"}, "validation": {"name": "validation", "num_bytes": 86021, "num_examples": 955, "dataset_name": " tweet_eval"}}, "download_checksums": {"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/irony/train_text.txt": {"num_bytes": 231594, "checksum": "a888125a44f7dfaa25b026318748d0e62cc9a300d20f66eafd62011a19eaea23"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/irony/train_labels.txt": {"num_bytes": 5724, "checksum": "fc69e6106c0f1f433a91536e08f83c71a391d7b219f7684d42f243a8089af77d"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/irony/test_text.txt": {"num_bytes": 68057, "checksum": "53103da934a7308eee82f05f2a9781a8ea3e88604fdc1e02d3101108505c64be"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/irony/test_labels.txt": {"num_bytes": 1568, "checksum": "08e2095e1725e74907a380614c220204e356bb46e3e8c93deb74e83e5b15ab38"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/irony/val_text.txt": {"num_bytes": 76760, "checksum": "8806cf3793e300a485cfae34892fc3a0a2f9a183deb06c750c6531515c83051e"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/irony/val_labels.txt": {"num_bytes": 1910, "checksum": "ccf429f63b4e8d0e7f425ca09445f7c31f7cea8a1b7c283b015b117c4002fd07"}}, "download_size": 385613, "post_processing_size": null, "dataset_size": 421113, "size_in_bytes": 806726}, "offensive": {"description": "TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.\n", "citation": "@inproceedings{barbieri2020tweeteval,\n title={{TweetEval:Unified Benchmark and Comparative Evaluation for Tweet Classification}},\n author={Barbieri, Francesco and Camacho-Collados, Jose and Espinosa-Anke, Luis and Neves, Leonardo},\n booktitle={Proceedings of Findings of EMNLP},\n year={2020}\n}\n", "homepage": "https://github.com/cardiffnlp/tweeteval", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 2, "names": ["non-offensive", "offensive"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "builder_name": " tweet_eval", "config_name": "offensive", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 1648069, "num_examples": 11916, "dataset_name": " tweet_eval"}, "test": {"name": "test", "num_bytes": 135477, "num_examples": 860, "dataset_name": " tweet_eval"}, "validation": {"name": "validation", "num_bytes": 192421, "num_examples": 1324, "dataset_name": " tweet_eval"}}, "download_checksums": {"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/offensive/train_text.txt": {"num_bytes": 1529074, "checksum": "78a7a32e38b10af7d8970b008bf17f661c8d0a90dad145fa0fa6a944669650db"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/offensive/train_labels.txt": {"num_bytes": 23832, "checksum": "c0b7d6ebdaa4ebcf6fc557ef1e775d92eda160218a0e3b1dd48eb8234dc892a6"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/offensive/test_text.txt": {"num_bytes": 126921, "checksum": "25b08c3333c26190f1023961c4508ec9aab24d4722b1a3ea7a6040724c120547"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/offensive/test_labels.txt": {"num_bytes": 1720, "checksum": "41d05a7aa0b01f5dafab21b95adb4f979cb4226c046ff315702774d10dac1605"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/offensive/val_text.txt": {"num_bytes": 179188, "checksum": "816f36d180c35f15a5104838cb73856a0bef42043482fe738f3481b06242a55c"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/offensive/val_labels.txt": {"num_bytes": 2648, "checksum": "ed2deb776bd1c52fb8221fadd3360e32d9dfe46842d78053528126e46363a258"}}, "download_size": 1863383, "post_processing_size": null, "dataset_size": 1975967, "size_in_bytes": 3839350}, "sentiment": {"description": "TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.\n", "citation": "@inproceedings{barbieri2020tweeteval,\n title={{TweetEval:Unified Benchmark and Comparative Evaluation for Tweet Classification}},\n author={Barbieri, Francesco and Camacho-Collados, Jose and Espinosa-Anke, Luis and Neves, Leonardo},\n booktitle={Proceedings of Findings of EMNLP},\n year={2020}\n}\n", "homepage": "https://github.com/cardiffnlp/tweeteval", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 3, "names": ["negative", "neutral", "positive"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "builder_name": " tweet_eval", "config_name": "sentiment", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 5425142, "num_examples": 45615, "dataset_name": " tweet_eval"}, "test": {"name": "test", "num_bytes": 1279548, "num_examples": 12284, "dataset_name": " tweet_eval"}, "validation": {"name": "validation", "num_bytes": 239088, "num_examples": 2000, "dataset_name": " tweet_eval"}}, "download_checksums": {"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/sentiment/train_text.txt": {"num_bytes": 4970029, "checksum": "368f01052ea6fd8ffc408a2a2e6ac9669e31542581a0396ef16591ea26eb98a6"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/sentiment/train_labels.txt": {"num_bytes": 91230, "checksum": "122bfb1732fb6995b0e5c5f726c0ba457c469c3b6e60513007ce5037f23e65d4"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/sentiment/test_text.txt": {"num_bytes": 1156877, "checksum": "09a93a55c63fd93f97485ef7302889d7edb4091cd49733aa37da094f0bfa0675"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/sentiment/test_labels.txt": {"num_bytes": 24568, "checksum": "6afb4afe9374d1f983bcf9a7c79b108d0f37fdf020a83f30488309bed215db9d"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/sentiment/val_text.txt": {"num_bytes": 219137, "checksum": "e5b021e6fc45064c260b09814b803d8f56cada519c4d952d72f43d48a350a964"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/sentiment/val_labels.txt": {"num_bytes": 4000, "checksum": "b4566926c72e2e4e2916c864def94e76c4cdde52446af2c7ba4fc2006e057e51"}}, "download_size": 6465841, "post_processing_size": null, "dataset_size": 6943778, "size_in_bytes": 13409619}, "stance_abortion": {"description": "TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.\n", "citation": "@inproceedings{barbieri2020tweeteval,\n title={{TweetEval:Unified Benchmark and Comparative Evaluation for Tweet Classification}},\n author={Barbieri, Francesco and Camacho-Collados, Jose and Espinosa-Anke, Luis and Neves, Leonardo},\n booktitle={Proceedings of Findings of EMNLP},\n year={2020}\n}\n", "homepage": "https://github.com/cardiffnlp/tweeteval", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 3, "names": ["none", "against", "favor"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "builder_name": " tweet_eval", "config_name": "stance_abortion", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 68698, "num_examples": 587, "dataset_name": " tweet_eval"}, "test": {"name": "test", "num_bytes": 33175, "num_examples": 280, "dataset_name": " tweet_eval"}, "validation": {"name": "validation", "num_bytes": 7661, "num_examples": 66, "dataset_name": " tweet_eval"}}, "download_checksums": {"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/abortion/train_text.txt": {"num_bytes": 62828, "checksum": "a421d5b8fd9f972970b9275b83f65745bf81986d2a412b4caa2ba071f3efa916"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/abortion/train_labels.txt": {"num_bytes": 1174, "checksum": "e6786a594bd9a083c524a0f420c690351140b52af288f487cb4772d29675b014"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/abortion/test_text.txt": {"num_bytes": 30371, "checksum": "bf0e16a0b8ca4cf0ab90efbc560db3151c288fc842f5e3c6554e8589d521556a"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/abortion/test_labels.txt": {"num_bytes": 560, "checksum": "c90e6d36d863f876d6661620d37b613b4b07858a5277c8d6623713ee59ca451c"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/abortion/val_text.txt": {"num_bytes": 6997, "checksum": "0428ab3f2894936f2445a9020763c2bd19ed42986872168bb65886dede5843fd"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/abortion/val_labels.txt": {"num_bytes": 132, "checksum": "8df57a50823d5f3683ecf75d824a42e3b08eb52e25e3e2d6928f523097a0c050"}}, "download_size": 102062, "post_processing_size": null, "dataset_size": 109534, "size_in_bytes": 211596}, "stance_atheism": {"description": "TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.\n", "citation": "@inproceedings{barbieri2020tweeteval,\n title={{TweetEval:Unified Benchmark and Comparative Evaluation for Tweet Classification}},\n author={Barbieri, Francesco and Camacho-Collados, Jose and Espinosa-Anke, Luis and Neves, Leonardo},\n booktitle={Proceedings of Findings of EMNLP},\n year={2020}\n}\n", "homepage": "https://github.com/cardiffnlp/tweeteval", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 3, "names": ["none", "against", "favor"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "builder_name": " tweet_eval", "config_name": "stance_atheism", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 54779, "num_examples": 461, "dataset_name": " tweet_eval"}, "test": {"name": "test", "num_bytes": 25720, "num_examples": 220, "dataset_name": " tweet_eval"}, "validation": {"name": "validation", "num_bytes": 6324, "num_examples": 52, "dataset_name": " tweet_eval"}}, "download_checksums": {"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/atheism/train_text.txt": {"num_bytes": 50165, "checksum": "0e82f1d4a16d79a38a68aee761762cf8a846bc8f7f9395670ca44e2ecf2f58f7"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/atheism/train_labels.txt": {"num_bytes": 922, "checksum": "a764aac1a75ccb32c4ffc4c03c66dc365cb50f013d3e94549bf775636cbc8373"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/atheism/test_text.txt": {"num_bytes": 23516, "checksum": "16c5336b2cba606ca63a6afcc50241be63a8fccf021628c6505449439b9d54b3"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/atheism/test_labels.txt": {"num_bytes": 440, "checksum": "4ef7c9398d265cfac625092c834e43cef9da9cb318e563493abb64f65dfe1b52"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/atheism/val_text.txt": {"num_bytes": 5800, "checksum": "5fe14c4c01f87a45dba640dddbb1d1909a893f9565f159c48fa1ba35bb46c209"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/atheism/val_labels.txt": {"num_bytes": 104, "checksum": "638095b3582f927fd1481cdb8d1f9f670f8d27880baf32c0b26c5946fd8f8292"}}, "download_size": 80947, "post_processing_size": null, "dataset_size": 86823, "size_in_bytes": 167770}, "stance_climate": {"description": "TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.\n", "citation": "@inproceedings{barbieri2020tweeteval,\n title={{TweetEval:Unified Benchmark and Comparative Evaluation for Tweet Classification}},\n author={Barbieri, Francesco and Camacho-Collados, Jose and Espinosa-Anke, Luis and Neves, Leonardo},\n booktitle={Proceedings of Findings of EMNLP},\n year={2020}\n}\n", "homepage": "https://github.com/cardiffnlp/tweeteval", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 3, "names": ["none", "against", "favor"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "builder_name": " tweet_eval", "config_name": "stance_climate", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 40253, "num_examples": 355, "dataset_name": " tweet_eval"}, "test": {"name": "test", "num_bytes": 19929, "num_examples": 169, "dataset_name": " tweet_eval"}, "validation": {"name": "validation", "num_bytes": 4805, "num_examples": 40, "dataset_name": " tweet_eval"}}, "download_checksums": {"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/climate/train_text.txt": {"num_bytes": 36699, "checksum": "4803211832d318026323a8e5014cff1b95e1c8c3854378101e5d1a8c82582eb7"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/climate/train_labels.txt": {"num_bytes": 710, "checksum": "d6274f55bc95f5a7f2ae591b886c1414a7664aaf4e0c609f4ba6cf377929af18"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/climate/test_text.txt": {"num_bytes": 18235, "checksum": "41ee8ee2ad3c36e0629654fdb271f37775197c79be8b299adbeadd2003b63c53"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/climate/test_labels.txt": {"num_bytes": 338, "checksum": "193c9f2358f61d9efe558324ec89ecaf08e600a44b68128f47838c01d9f98dfd"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/climate/val_text.txt": {"num_bytes": 4401, "checksum": "fc5714703add266801ee2fd98296ea20ec0879e89cdb9f906d9812d9f640f2ba"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/climate/val_labels.txt": {"num_bytes": 80, "checksum": "0cb133ab9b137292f075210db45f7e293dc52798a4e21e59037bfcfe66c97aa6"}}, "download_size": 60463, "post_processing_size": null, "dataset_size": 64987, "size_in_bytes": 125450}, "stance_feminist": {"description": "TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.\n", "citation": "@inproceedings{barbieri2020tweeteval,\n title={{TweetEval:Unified Benchmark and Comparative Evaluation for Tweet Classification}},\n author={Barbieri, Francesco and Camacho-Collados, Jose and Espinosa-Anke, Luis and Neves, Leonardo},\n booktitle={Proceedings of Findings of EMNLP},\n year={2020}\n}\n", "homepage": "https://github.com/cardiffnlp/tweeteval", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 3, "names": ["none", "against", "favor"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "builder_name": " tweet_eval", "config_name": "stance_feminist", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 70513, "num_examples": 597, "dataset_name": " tweet_eval"}, "test": {"name": "test", "num_bytes": 33309, "num_examples": 285, "dataset_name": " tweet_eval"}, "validation": {"name": "validation", "num_bytes": 8039, "num_examples": 67, "dataset_name": " tweet_eval"}}, "download_checksums": {"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/feminist/train_text.txt": {"num_bytes": 64539, "checksum": "c176e6663973c8e78bfa92ba1e8874a70cc5358567d71584a90943bc6525eaab"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/feminist/train_labels.txt": {"num_bytes": 1194, "checksum": "abd4f196d801423bb0daba8c0ecf5b3efba1f10e8f410c3dfa360b50c8b9c685"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/feminist/test_text.txt": {"num_bytes": 30455, "checksum": "1bfdbdc2af64fd62dcc775d1288e192ac8ff805ef27ccf3aaac54a98616eefda"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/feminist/test_labels.txt": {"num_bytes": 570, "checksum": "ddbde6d253ee47c5d5ef8bc5386270fde45cf088d3be70bba9c382b8a024897a"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/feminist/val_text.txt": {"num_bytes": 7365, "checksum": "3518b2ddcf696626a7243d7cea720a975718c7a52a5a086931be87897c1de58b"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/feminist/val_labels.txt": {"num_bytes": 134, "checksum": "399e0d468d0e4ead7a445f69efdf35876c835acf4cefc00a16f451a5d42e5c13"}}, "download_size": 104257, "post_processing_size": null, "dataset_size": 111861, "size_in_bytes": 216118}, "stance_hillary": {"description": "TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.\n", "citation": "@inproceedings{barbieri2020tweeteval,\n title={{TweetEval:Unified Benchmark and Comparative Evaluation for Tweet Classification}},\n author={Barbieri, Francesco and Camacho-Collados, Jose and Espinosa-Anke, Luis and Neves, Leonardo},\n booktitle={Proceedings of Findings of EMNLP},\n year={2020}\n}\n", "homepage": "https://github.com/cardiffnlp/tweeteval", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 3, "names": ["none", "against", "favor"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "builder_name": " tweet_eval", "config_name": "stance_hillary", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 69600, "num_examples": 620, "dataset_name": " tweet_eval"}, "test": {"name": "test", "num_bytes": 34491, "num_examples": 295, "dataset_name": " tweet_eval"}, "validation": {"name": "validation", "num_bytes": 7536, "num_examples": 69, "dataset_name": " tweet_eval"}}, "download_checksums": {"https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/hillary/train_text.txt": {"num_bytes": 63398, "checksum": "0bd735de895cb74d63c224e64e3d955cac99be97aa225f803fe4d2f5978a2c99"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/hillary/train_labels.txt": {"num_bytes": 1240, "checksum": "0ea5753d13a717a9e91581d1d89c0b5206c8f905f0a717b2b27d02dbf419250d"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/hillary/test_text.txt": {"num_bytes": 31537, "checksum": "5c4e020285a62cfd88f264849e1db242ded356c171b1a68dd0050b76635053aa"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/hillary/test_labels.txt": {"num_bytes": 590, "checksum": "068468f6a72b85dfb65bf10e45f2453fa082d1ea9d7a40e7f560d5b6d75027f3"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/hillary/val_text.txt": {"num_bytes": 6842, "checksum": "9714b7dcc8617e095433d7b63df8aa155eb84216b9ac9195105ab83d85cd248d"}, "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/stance/hillary/val_labels.txt": {"num_bytes": 138, "checksum": "e5d44c771b7349a4a74309f56ca072fdf8f1c015068d519ca2ed3a931c833606"}}, "download_size": 103745, "post_processing_size": null, "dataset_size": 111627, "size_in_bytes": 215372}}
|
|
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:95e248466f9153ba8ee6f990f0ad26c7d77520326d77e697aaf7e09df350e65d
|
3 |
+
size 3047341
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d4f587c86b7459b727ae1c84d39ca7fd42631e51d43cc3f95242dcd4c696077a
|
3 |
+
size 2609973
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:48f3d0140d19153a1a14e2940d62b63dcdfb24c4b958c65bd9ea1dfb777f391a
|
3 |
+
size 281994
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:460be5e853577cad9e070549a3bb0eaecccfa277e07db69873fc8f46bd57299f
|
3 |
+
size 105421
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:448fb3d7b94184ca25652d7361d82bbc50c20cd0d10a984ce3596a1aa7e1d75c
|
3 |
+
size 233004
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:20e78c12dc5a17c2011a708cc838b8b0251b87f624f71dbdd354a2e3857aab73
|
3 |
+
size 28591
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c07de83b04e013b956c66187e1fd454fadb6f4caacc23e11f5fb3f80ddc2abd2
|
3 |
+
size 277846
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:da2c391c2050df2d90d3d9e7b8bca48a18056293138b5c6f08c7bf61c5b84668
|
3 |
+
size 815722
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f7307944af6e873feb693a9c0853439af8d0543e571ae8a2c4cfc91e01016c68
|
3 |
+
size 102778
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d6bcf4eae70637cacc9239773684a37cbfa3ef98fde5f173355651afee477862
|
3 |
+
size 53984
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:541de8c14f53024660fa5ae1534489b51106c5f4905da294e47ded7e09d73ad0
|
3 |
+
size 182571
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2ac02a5943b02808f1998dfacba7fce89033be6b3e4bbb465a26faafabf21191
|
3 |
+
size 61092
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:15b87adaf1850d8e3e104be6240255813d48d776940be546ad353fee9fcfb099
|
3 |
+
size 93731
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4c2cb443cb5b956aa8578a6e18b2f9e50ce43217774e3cf171d53e16a1db81a3
|
3 |
+
size 1019132
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1161b4e6cc7cbdaa667d4fc131f9be0ea3db755616126329dda7eb521cd88c36
|
3 |
+
size 121665
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f73aea52d43879223019e733b944780f7e9203ee32459ea006f554011379fcdf
|
3 |
+
size 900553
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:884ed7990ba4058aa78372f5f9590db308c393820545d9c4f2363f5a2ea12fbf
|
3 |
+
size 3781982
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba8e7513047d4991bcdcc0a22aaa70777e47b609f1b739da5117baf7893f3e56
|
3 |
+
size 167140
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4c25f9b438767cbc25f143beac32f7ade213baa5c436be32733e0b718d1140f6
|
3 |
+
size 22513
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5d90ce50ba5ed14d7600bc9ca919c2ac2b9f58141b80d63b96fa17027607bd20
|
3 |
+
size 43712
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f8ac5a9696577bb1622b12bec44ac8222f26f2109cb6a2c0abc27305fa93bf56
|
3 |
+
size 7292
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b92549ee5a2d90448a1fac4ed509765e1b166abb91705e997a62a54e3606127c
|
3 |
+
size 19358
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4d915d95b5a313db91f3737d095aa89db6b9e749c7838e00fe844ff5a6f98661
|
3 |
+
size 36467
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f10c1952169a56231bf2f4c9548e98843219311170fec4fe1fd4557cc1b85f9c
|
3 |
+
size 6440
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cda4111271a268e0d2513d8effa704395f59ce2b8ece90f93e4a96d2c16dc8b9
|
3 |
+
size 14897
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:248ba3695bba19f1fcef9c65421d81c65fa3a76570f1d7d3766872022e96b224
|
3 |
+
size 28127
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4f4d174e3f9f7c526b532409fa4ca59abd27d9eb35c03df4daa2e9e64366f8c5
|
3 |
+
size 5469
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2230a3e35acd62ace8d77aa6fd8731a724833d540f732a75fb6327f2cdf1e0ea
|
3 |
+
size 23416
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f6ae1f6ee44c1e0ac0523bab6278febf8a62387305c24aad3ea9701868b81e1a
|
3 |
+
size 45301
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4bc15f3d903ba99da50a200e211e1506c8117378310f2fd54edb69d7d5d70493
|
3 |
+
size 7628
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:62f0fcbf8356ce7df510365dae39c2928b7edf75e0576c09f5196f8483315d2a
|
3 |
+
size 23516
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:61b7730fdffe10728554b85b3d36af454df6185fcebf214697215965a28dc364
|
3 |
+
size 43297
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:366e5594562feaadbfddf307b8d5940461fff70efd7654bcbee9dde0d8c32068
|
3 |
+
size 7244
|
@@ -1,249 +0,0 @@
|
|
1 |
-
# coding=utf-8
|
2 |
-
# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
|
3 |
-
#
|
4 |
-
# Licensed under the Apache License, Version 2.0 (the "License");
|
5 |
-
# you may not use this file except in compliance with the License.
|
6 |
-
# You may obtain a copy of the License at
|
7 |
-
#
|
8 |
-
# http://www.apache.org/licenses/LICENSE-2.0
|
9 |
-
#
|
10 |
-
# Unless required by applicable law or agreed to in writing, software
|
11 |
-
# distributed under the License is distributed on an "AS IS" BASIS,
|
12 |
-
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
13 |
-
# See the License for the specific language governing permissions and
|
14 |
-
# limitations under the License.
|
15 |
-
"""The Tweet Eval Datasets"""
|
16 |
-
|
17 |
-
|
18 |
-
import datasets
|
19 |
-
|
20 |
-
|
21 |
-
_CITATION = """\
|
22 |
-
@inproceedings{barbieri2020tweeteval,
|
23 |
-
title={{TweetEval:Unified Benchmark and Comparative Evaluation for Tweet Classification}},
|
24 |
-
author={Barbieri, Francesco and Camacho-Collados, Jose and Espinosa-Anke, Luis and Neves, Leonardo},
|
25 |
-
booktitle={Proceedings of Findings of EMNLP},
|
26 |
-
year={2020}
|
27 |
-
}
|
28 |
-
"""
|
29 |
-
|
30 |
-
_DESCRIPTION = """\
|
31 |
-
TweetEval consists of seven heterogenous tasks in Twitter, all framed as multi-class tweet classification. All tasks have been unified into the same benchmark, with each dataset presented in the same format and with fixed training, validation and test splits.
|
32 |
-
"""
|
33 |
-
|
34 |
-
_HOMEPAGE = "https://github.com/cardiffnlp/tweeteval"
|
35 |
-
|
36 |
-
_LICENSE = ""
|
37 |
-
|
38 |
-
URL = "https://raw.githubusercontent.com/cardiffnlp/tweeteval/main/datasets/"
|
39 |
-
|
40 |
-
_URLs = {
|
41 |
-
"emoji": {
|
42 |
-
"train_text": URL + "emoji/train_text.txt",
|
43 |
-
"train_labels": URL + "emoji/train_labels.txt",
|
44 |
-
"test_text": URL + "emoji/test_text.txt",
|
45 |
-
"test_labels": URL + "emoji/test_labels.txt",
|
46 |
-
"val_text": URL + "emoji/val_text.txt",
|
47 |
-
"val_labels": URL + "emoji/val_labels.txt",
|
48 |
-
},
|
49 |
-
"emotion": {
|
50 |
-
"train_text": URL + "emotion/train_text.txt",
|
51 |
-
"train_labels": URL + "emotion/train_labels.txt",
|
52 |
-
"test_text": URL + "emotion/test_text.txt",
|
53 |
-
"test_labels": URL + "emotion/test_labels.txt",
|
54 |
-
"val_text": URL + "emotion/val_text.txt",
|
55 |
-
"val_labels": URL + "emotion/val_labels.txt",
|
56 |
-
},
|
57 |
-
"hate": {
|
58 |
-
"train_text": URL + "hate/train_text.txt",
|
59 |
-
"train_labels": URL + "hate/train_labels.txt",
|
60 |
-
"test_text": URL + "hate/test_text.txt",
|
61 |
-
"test_labels": URL + "hate/test_labels.txt",
|
62 |
-
"val_text": URL + "hate/val_text.txt",
|
63 |
-
"val_labels": URL + "hate/val_labels.txt",
|
64 |
-
},
|
65 |
-
"irony": {
|
66 |
-
"train_text": URL + "irony/train_text.txt",
|
67 |
-
"train_labels": URL + "irony/train_labels.txt",
|
68 |
-
"test_text": URL + "irony/test_text.txt",
|
69 |
-
"test_labels": URL + "irony/test_labels.txt",
|
70 |
-
"val_text": URL + "irony/val_text.txt",
|
71 |
-
"val_labels": URL + "irony/val_labels.txt",
|
72 |
-
},
|
73 |
-
"offensive": {
|
74 |
-
"train_text": URL + "offensive/train_text.txt",
|
75 |
-
"train_labels": URL + "offensive/train_labels.txt",
|
76 |
-
"test_text": URL + "offensive/test_text.txt",
|
77 |
-
"test_labels": URL + "offensive/test_labels.txt",
|
78 |
-
"val_text": URL + "offensive/val_text.txt",
|
79 |
-
"val_labels": URL + "offensive/val_labels.txt",
|
80 |
-
},
|
81 |
-
"sentiment": {
|
82 |
-
"train_text": URL + "sentiment/train_text.txt",
|
83 |
-
"train_labels": URL + "sentiment/train_labels.txt",
|
84 |
-
"test_text": URL + "sentiment/test_text.txt",
|
85 |
-
"test_labels": URL + "sentiment/test_labels.txt",
|
86 |
-
"val_text": URL + "sentiment/val_text.txt",
|
87 |
-
"val_labels": URL + "sentiment/val_labels.txt",
|
88 |
-
},
|
89 |
-
"stance": {
|
90 |
-
"abortion": {
|
91 |
-
"train_text": URL + "stance/abortion/train_text.txt",
|
92 |
-
"train_labels": URL + "stance/abortion/train_labels.txt",
|
93 |
-
"test_text": URL + "stance/abortion/test_text.txt",
|
94 |
-
"test_labels": URL + "stance/abortion/test_labels.txt",
|
95 |
-
"val_text": URL + "stance/abortion/val_text.txt",
|
96 |
-
"val_labels": URL + "stance/abortion/val_labels.txt",
|
97 |
-
},
|
98 |
-
"atheism": {
|
99 |
-
"train_text": URL + "stance/atheism/train_text.txt",
|
100 |
-
"train_labels": URL + "stance/atheism/train_labels.txt",
|
101 |
-
"test_text": URL + "stance/atheism/test_text.txt",
|
102 |
-
"test_labels": URL + "stance/atheism/test_labels.txt",
|
103 |
-
"val_text": URL + "stance/atheism/val_text.txt",
|
104 |
-
"val_labels": URL + "stance/atheism/val_labels.txt",
|
105 |
-
},
|
106 |
-
"climate": {
|
107 |
-
"train_text": URL + "stance/climate/train_text.txt",
|
108 |
-
"train_labels": URL + "stance/climate/train_labels.txt",
|
109 |
-
"test_text": URL + "stance/climate/test_text.txt",
|
110 |
-
"test_labels": URL + "stance/climate/test_labels.txt",
|
111 |
-
"val_text": URL + "stance/climate/val_text.txt",
|
112 |
-
"val_labels": URL + "stance/climate/val_labels.txt",
|
113 |
-
},
|
114 |
-
"feminist": {
|
115 |
-
"train_text": URL + "stance/feminist/train_text.txt",
|
116 |
-
"train_labels": URL + "stance/feminist/train_labels.txt",
|
117 |
-
"test_text": URL + "stance/feminist/test_text.txt",
|
118 |
-
"test_labels": URL + "stance/feminist/test_labels.txt",
|
119 |
-
"val_text": URL + "stance/feminist/val_text.txt",
|
120 |
-
"val_labels": URL + "stance/feminist/val_labels.txt",
|
121 |
-
},
|
122 |
-
"hillary": {
|
123 |
-
"train_text": URL + "stance/hillary/train_text.txt",
|
124 |
-
"train_labels": URL + "stance/hillary/train_labels.txt",
|
125 |
-
"test_text": URL + "stance/hillary/test_text.txt",
|
126 |
-
"test_labels": URL + "stance/hillary/test_labels.txt",
|
127 |
-
"val_text": URL + "stance/hillary/val_text.txt",
|
128 |
-
"val_labels": URL + "stance/hillary/val_labels.txt",
|
129 |
-
},
|
130 |
-
},
|
131 |
-
}
|
132 |
-
|
133 |
-
|
134 |
-
class TweetEvalConfig(datasets.BuilderConfig):
|
135 |
-
def __init__(self, *args, type=None, sub_type=None, **kwargs):
|
136 |
-
super().__init__(
|
137 |
-
*args,
|
138 |
-
name=f"{type}" if type != "stance" else f"{type}_{sub_type}",
|
139 |
-
**kwargs,
|
140 |
-
)
|
141 |
-
self.type = type
|
142 |
-
self.sub_type = sub_type
|
143 |
-
|
144 |
-
|
145 |
-
class TweetEval(datasets.GeneratorBasedBuilder):
|
146 |
-
"""TweetEval Dataset."""
|
147 |
-
|
148 |
-
BUILDER_CONFIGS = [
|
149 |
-
TweetEvalConfig(
|
150 |
-
type=key,
|
151 |
-
sub_type=None,
|
152 |
-
version=datasets.Version("1.1.0"),
|
153 |
-
description=f"This part of my dataset covers {key} part of TweetEval Dataset.",
|
154 |
-
)
|
155 |
-
for key in list(_URLs.keys())
|
156 |
-
if key != "stance"
|
157 |
-
] + [
|
158 |
-
TweetEvalConfig(
|
159 |
-
type="stance",
|
160 |
-
sub_type=key,
|
161 |
-
version=datasets.Version("1.1.0"),
|
162 |
-
description=f"This part of my dataset covers stance_{key} part of TweetEval Dataset.",
|
163 |
-
)
|
164 |
-
for key in list(_URLs["stance"].keys())
|
165 |
-
]
|
166 |
-
|
167 |
-
def _info(self):
|
168 |
-
if self.config.type == "stance":
|
169 |
-
names = ["none", "against", "favor"]
|
170 |
-
elif self.config.type == "sentiment":
|
171 |
-
names = ["negative", "neutral", "positive"]
|
172 |
-
elif self.config.type == "offensive":
|
173 |
-
names = ["non-offensive", "offensive"]
|
174 |
-
elif self.config.type == "irony":
|
175 |
-
names = ["non_irony", "irony"]
|
176 |
-
elif self.config.type == "hate":
|
177 |
-
names = ["non-hate", "hate"]
|
178 |
-
elif self.config.type == "emoji":
|
179 |
-
names = [
|
180 |
-
"β€",
|
181 |
-
"π",
|
182 |
-
"π",
|
183 |
-
"π",
|
184 |
-
"π₯",
|
185 |
-
"π",
|
186 |
-
"π",
|
187 |
-
"β¨",
|
188 |
-
"π",
|
189 |
-
"π",
|
190 |
-
"π·",
|
191 |
-
"πΊπΈ",
|
192 |
-
"β",
|
193 |
-
"π",
|
194 |
-
"π",
|
195 |
-
"π―",
|
196 |
-
"π",
|
197 |
-
"π",
|
198 |
-
"πΈ",
|
199 |
-
"π",
|
200 |
-
]
|
201 |
-
|
202 |
-
else:
|
203 |
-
names = ["anger", "joy", "optimism", "sadness"]
|
204 |
-
|
205 |
-
return datasets.DatasetInfo(
|
206 |
-
description=_DESCRIPTION,
|
207 |
-
features=datasets.Features(
|
208 |
-
{"text": datasets.Value("string"), "label": datasets.features.ClassLabel(names=names)}
|
209 |
-
),
|
210 |
-
supervised_keys=None,
|
211 |
-
homepage=_HOMEPAGE,
|
212 |
-
license=_LICENSE,
|
213 |
-
citation=_CITATION,
|
214 |
-
)
|
215 |
-
|
216 |
-
def _split_generators(self, dl_manager):
|
217 |
-
"""Returns SplitGenerators."""
|
218 |
-
if self.config.type != "stance":
|
219 |
-
my_urls = _URLs[self.config.type]
|
220 |
-
else:
|
221 |
-
my_urls = _URLs[self.config.type][self.config.sub_type]
|
222 |
-
data_dir = dl_manager.download_and_extract(my_urls)
|
223 |
-
return [
|
224 |
-
datasets.SplitGenerator(
|
225 |
-
name=datasets.Split.TRAIN,
|
226 |
-
# These kwargs will be passed to _generate_examples
|
227 |
-
gen_kwargs={"text_path": data_dir["train_text"], "labels_path": data_dir["train_labels"]},
|
228 |
-
),
|
229 |
-
datasets.SplitGenerator(
|
230 |
-
name=datasets.Split.TEST,
|
231 |
-
# These kwargs will be passed to _generate_examples
|
232 |
-
gen_kwargs={"text_path": data_dir["test_text"], "labels_path": data_dir["test_labels"]},
|
233 |
-
),
|
234 |
-
datasets.SplitGenerator(
|
235 |
-
name=datasets.Split.VALIDATION,
|
236 |
-
# These kwargs will be passed to _generate_examples
|
237 |
-
gen_kwargs={"text_path": data_dir["val_text"], "labels_path": data_dir["val_labels"]},
|
238 |
-
),
|
239 |
-
]
|
240 |
-
|
241 |
-
def _generate_examples(self, text_path, labels_path):
|
242 |
-
"""Yields examples."""
|
243 |
-
|
244 |
-
with open(text_path, encoding="utf-8") as f:
|
245 |
-
texts = f.readlines()
|
246 |
-
with open(labels_path, encoding="utf-8") as f:
|
247 |
-
labels = f.readlines()
|
248 |
-
for i, text in enumerate(texts):
|
249 |
-
yield i, {"text": text.strip(), "label": int(labels[i].strip())}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|