ArunaSaraswathy commited on
Commit
b772df8
1 Parent(s): 7a21057

Training complete

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "distilbert-base-cased",
3
  "activation": "gelu",
4
  "architectures": [
5
  "DistilBertForTokenClassification"
 
1
  {
2
+ "_name_or_path": "ArunaSaraswathy/my_model",
3
  "activation": "gelu",
4
  "architectures": [
5
  "DistilBertForTokenClassification"
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4723157c16496afcaa7dca14099dc7a985d23cd55c822ad6a7c88835f0224b39
3
  size 260824741
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56907f0ae5a807d6f7dbb2ec5cc6d31474fa96186bdd49fa9e83b9b066d470b4
3
  size 260824741
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": false,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "name_or_path": "distilbert-base-cased",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "DistilBertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff