alvarodt commited on
Commit
f0a0fda
·
1 Parent(s): 3976d9f

Training in progress, step 500

Browse files
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
config.json CHANGED
@@ -1,7 +1,8 @@
1
  {
 
2
  "activation": "gelu",
3
  "architectures": [
4
- "DistilBertModel"
5
  ],
6
  "attention_dropout": 0.1,
7
  "dim": 768,
@@ -11,12 +12,13 @@
11
  "max_position_embeddings": 512,
12
  "model_type": "distilbert",
13
  "n_heads": 12,
14
- "n_layers": 4,
15
  "pad_token_id": 0,
16
  "qa_dropout": 0.1,
17
  "seq_classif_dropout": 0.2,
18
  "sinusoidal_pos_embds": false,
 
19
  "torch_dtype": "float32",
20
- "transformers_version": "4.26.0",
21
  "vocab_size": 30522
22
  }
 
1
  {
2
+ "_name_or_path": "distilbert-base-uncased",
3
  "activation": "gelu",
4
  "architectures": [
5
+ "DistilBertForMaskedLM"
6
  ],
7
  "attention_dropout": 0.1,
8
  "dim": 768,
 
12
  "max_position_embeddings": 512,
13
  "model_type": "distilbert",
14
  "n_heads": 12,
15
+ "n_layers": 6,
16
  "pad_token_id": 0,
17
  "qa_dropout": 0.1,
18
  "seq_classif_dropout": 0.2,
19
  "sinusoidal_pos_embds": false,
20
+ "tie_weights_": true,
21
  "torch_dtype": "float32",
22
+ "transformers_version": "4.26.1",
23
  "vocab_size": 30522
24
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4c7232dc79925f1bd1219be3c340eb1f8d69b86be7f973304aa143d63f0b29c6
3
- size 208771005
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f862edf708eebc7fcdf6e43c2aedc3421bd640d8f9b1bee656674f4ceded496
3
+ size 267976561
special_tokens_map.json CHANGED
@@ -1,9 +1,7 @@
1
  {
2
- "bos_token": "<|BOS|>",
3
- "cls_token": "<|CLS|>",
4
- "eos_token": "<|EOS|>",
5
- "mask_token": "<|MASK|>",
6
- "pad_token": "<|PAD|>",
7
- "sep_token": "<|SEP|>",
8
- "unk_token": "<|UNK|>"
9
  }
 
1
  {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
 
 
7
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
- "add_prefix_space": false,
3
- "bos_token": "<|BOS|>",
4
- "cls_token": "<|CLS|>",
5
- "eos_token": "<|EOS|>",
6
- "mask_token": "<|MASK|>",
7
- "model_max_length": 1024,
8
- "name_or_path": "gpt2",
9
- "pad_token": "<|PAD|>",
10
- "sep_token": "<|SEP|>",
11
  "special_tokens_map_file": null,
12
- "tokenizer_class": "GPT2Tokenizer",
13
- "unk_token": "<|UNK|>"
 
 
14
  }
 
1
  {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": true,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 66.0,
6
+ "name_or_path": "distilbert-base-uncased",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
 
 
9
  "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "DistilBertTokenizer",
13
+ "unk_token": "[UNK]"
14
  }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e3fa84723cc899fae0aa070f0eb821021ae4e9142ab251021c5a20073904b52
3
+ size 3515