autotrain-data-processor commited on
Commit
7e01043
·
1 Parent(s): 70cdc6a

Processed data from AutoTrain data processor ([2023-02-10 11:17 ]

Browse files
README.md ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ task_categories:
5
+ - text-classification
6
+
7
+ ---
8
+ # AutoTrain Dataset for project: dataset-mentions
9
+
10
+ ## Dataset Description
11
+
12
+ This dataset has been automatically processed by AutoTrain for project dataset-mentions.
13
+
14
+ ### Languages
15
+
16
+ The BCP-47 code for the dataset's language is en.
17
+
18
+ ## Dataset Structure
19
+
20
+ ### Data Instances
21
+
22
+ A sample from this dataset looks as follows:
23
+
24
+ ```json
25
+ [
26
+ {
27
+ "text": " How to use ```python from transformers import AutoTokenizer, AutoModel tokenizer = AutoTokenizer.from_pretrained(\"Geotrend/bert-base-en-fr-zh-ja-vi-cased\") model = AutoModel.from_pretrained(\"Geotrend/bert-base-en-fr-zh-ja-vi-cased\") ``` To generate other smaller versions of multilingual transformers please visit [our Github repo](https://github.com/Geotrend-research/smaller-transformers). ",
28
+ "target": 0
29
+ },
30
+ {
31
+ "text": " Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 24 - eval_batch_size: 24 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ",
32
+ "target": 1
33
+ }
34
+ ]
35
+ ```
36
+
37
+ ### Dataset Fields
38
+
39
+ The dataset has the following fields (also called "features"):
40
+
41
+ ```json
42
+ {
43
+ "text": "Value(dtype='string', id=None)",
44
+ "target": "ClassLabel(names=['dataset_mention', 'no_dataset_mention'], id=None)"
45
+ }
46
+ ```
47
+
48
+ ### Dataset Splits
49
+
50
+ This dataset is split into a train and validation split. The split sizes are as follow:
51
+
52
+ | Split name | Num samples |
53
+ | ------------ | ------------------- |
54
+ | train | 7428 |
55
+ | valid | 1858 |
processed/dataset_dict.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"splits": ["train", "valid"]}
processed/train/dataset.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0294723314f1e6bb4d9186ab4db08fdcf47b86805716bedb4ecfb04fc6b442c
3
+ size 2939072
processed/train/dataset_info.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "citation": "",
3
+ "description": "AutoTrain generated dataset",
4
+ "features": {
5
+ "text": {
6
+ "dtype": "string",
7
+ "_type": "Value"
8
+ },
9
+ "target": {
10
+ "names": [
11
+ "dataset_mention",
12
+ "no_dataset_mention"
13
+ ],
14
+ "_type": "ClassLabel"
15
+ }
16
+ },
17
+ "homepage": "",
18
+ "license": "",
19
+ "splits": {
20
+ "train": {
21
+ "name": "train",
22
+ "num_bytes": 2936898,
23
+ "num_examples": 7428,
24
+ "dataset_name": null
25
+ }
26
+ }
27
+ }
processed/train/state.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "dataset.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "0e021e6946da352a",
8
+ "_format_columns": [
9
+ "target",
10
+ "text"
11
+ ],
12
+ "_format_kwargs": {},
13
+ "_format_type": null,
14
+ "_indexes": {},
15
+ "_output_all_columns": false,
16
+ "_split": null
17
+ }
processed/valid/dataset.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6027215ac2c965bac4bc1af3a9063542c1a89607e3aaa636795b9c081947754f
3
+ size 905376
processed/valid/dataset_info.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "citation": "",
3
+ "description": "AutoTrain generated dataset",
4
+ "features": {
5
+ "text": {
6
+ "dtype": "string",
7
+ "_type": "Value"
8
+ },
9
+ "target": {
10
+ "names": [
11
+ "dataset_mention",
12
+ "no_dataset_mention"
13
+ ],
14
+ "_type": "ClassLabel"
15
+ }
16
+ },
17
+ "homepage": "",
18
+ "license": "",
19
+ "splits": {
20
+ "valid": {
21
+ "name": "valid",
22
+ "num_bytes": 904533,
23
+ "num_examples": 1858,
24
+ "dataset_name": null
25
+ }
26
+ }
27
+ }
processed/valid/state.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "dataset.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "01dc7a252529f9cf",
8
+ "_format_columns": [
9
+ "target",
10
+ "text"
11
+ ],
12
+ "_format_kwargs": {},
13
+ "_format_type": null,
14
+ "_indexes": {},
15
+ "_output_all_columns": false,
16
+ "_split": null
17
+ }