channudam commited on
Commit
d599b98
·
verified ·
1 Parent(s): 35f577a

End of training

Browse files
Files changed (4) hide show
  1. README.md +11 -11
  2. config.json +1 -1
  3. model.safetensors +1 -1
  4. tokenizer_config.json +7 -0
README.md CHANGED
@@ -9,13 +9,13 @@ model-index:
9
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
  should probably proofread and complete it, then remove this comment. -->
11
 
12
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/rayranger/huggingface/runs/9t8e6qv5)
13
  # khmer-trocr-base-printed
14
 
15
  This model was trained from scratch on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
- - Loss: 0.0783
18
- - Cer: 0.5191
19
 
20
  ## Model description
21
 
@@ -47,14 +47,14 @@ The following hyperparameters were used during training:
47
 
48
  | Training Loss | Epoch | Step | Validation Loss | Cer |
49
  |:-------------:|:------:|:----:|:---------------:|:------:|
50
- | 0.2278 | 0.4876 | 1000 | 0.2244 | 0.6099 |
51
- | 0.1517 | 0.9751 | 2000 | 0.1667 | 0.5711 |
52
- | 0.1147 | 1.4627 | 3000 | 0.1412 | 0.5578 |
53
- | 0.0927 | 1.9503 | 4000 | 0.1159 | 0.5552 |
54
- | 0.0669 | 2.4378 | 5000 | 0.1028 | 0.5458 |
55
- | 0.0551 | 2.9254 | 6000 | 0.0901 | 0.5270 |
56
- | 0.0304 | 3.4130 | 7000 | 0.0833 | 0.5170 |
57
- | 0.0182 | 3.9005 | 8000 | 0.0783 | 0.5191 |
58
 
59
 
60
  ### Framework versions
 
9
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
  should probably proofread and complete it, then remove this comment. -->
11
 
12
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/rayranger/huggingface/runs/q0t4izjn)
13
  # khmer-trocr-base-printed
14
 
15
  This model was trained from scratch on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Loss: 0.0635
18
+ - Cer: 0.5175
19
 
20
  ## Model description
21
 
 
47
 
48
  | Training Loss | Epoch | Step | Validation Loss | Cer |
49
  |:-------------:|:------:|:----:|:---------------:|:------:|
50
+ | 0.0662 | 0.4876 | 1000 | 0.1256 | 0.5637 |
51
+ | 0.0435 | 0.9751 | 2000 | 0.1191 | 0.5692 |
52
+ | 0.0412 | 1.4627 | 3000 | 0.0978 | 0.5385 |
53
+ | 0.0275 | 1.9503 | 4000 | 0.0896 | 0.5530 |
54
+ | 0.0212 | 2.4378 | 5000 | 0.0825 | 0.5518 |
55
+ | 0.0155 | 2.9254 | 6000 | 0.0732 | 0.5297 |
56
+ | 0.0084 | 3.4130 | 7000 | 0.0676 | 0.5107 |
57
+ | 0.005 | 3.9005 | 8000 | 0.0635 | 0.5175 |
58
 
59
 
60
  ### Framework versions
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/kaggle/input/khmerocr-trocr-base/KhmerOCR_TrOCR",
3
  "architectures": [
4
  "VisionEncoderDecoderModel"
5
  ],
 
1
  {
2
+ "_name_or_path": "/kaggle/input/khmerocr-trocr-fine-tune-v1/KhmerOCR_TrOCR",
3
  "architectures": [
4
  "VisionEncoderDecoderModel"
5
  ],
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:480b91120705e08bbee9571fe7f7180c7342f31264384a18b789f35fdcbc7b88
3
  size 608335608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a83faf51a7f8ba96f4a9a47b864002fbeab6383f4db8c2ff2601c0ae77de5f30
3
  size 608335608
tokenizer_config.json CHANGED
@@ -49,11 +49,18 @@
49
  "errors": "replace",
50
  "mask_token": "<mask>",
51
  "max_len": 150,
 
52
  "model_max_length": 150,
 
53
  "pad_token": "<pad>",
 
 
54
  "processor_class": "TrOCRProcessor",
55
  "sep_token": "</s>",
 
56
  "tokenizer_class": "RobertaTokenizer",
57
  "trim_offsets": true,
 
 
58
  "unk_token": "<unk>"
59
  }
 
49
  "errors": "replace",
50
  "mask_token": "<mask>",
51
  "max_len": 150,
52
+ "max_length": 145,
53
  "model_max_length": 150,
54
+ "pad_to_multiple_of": null,
55
  "pad_token": "<pad>",
56
+ "pad_token_type_id": 0,
57
+ "padding_side": "right",
58
  "processor_class": "TrOCRProcessor",
59
  "sep_token": "</s>",
60
+ "stride": 0,
61
  "tokenizer_class": "RobertaTokenizer",
62
  "trim_offsets": true,
63
+ "truncation_side": "right",
64
+ "truncation_strategy": "longest_first",
65
  "unk_token": "<unk>"
66
  }