Chung-Fan commited on
Commit
c94ee21
1 Parent(s): 6d75584

Training done for led-pubmed-20k

Browse files
README.md CHANGED
@@ -15,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [pszemraj/led-base-book-summary](https://huggingface.co/pszemraj/led-base-book-summary) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 1.2432
19
 
20
  ## Model description
21
 
@@ -47,14 +47,14 @@ The following hyperparameters were used during training:
47
 
48
  ### Training results
49
 
50
- | Training Loss | Epoch | Step | Validation Loss |
51
- |:-------------:|:-----:|:----:|:---------------:|
52
- | 1.2934 | 0.75 | 500 | 1.2432 |
53
 
54
 
55
  ### Framework versions
56
 
57
- - Transformers 4.38.2
58
  - Pytorch 2.2.1+cu121
59
- - Datasets 2.18.0
60
- - Tokenizers 0.15.2
 
15
 
16
  This model is a fine-tuned version of [pszemraj/led-base-book-summary](https://huggingface.co/pszemraj/led-base-book-summary) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 1.1066
19
 
20
  ## Model description
21
 
 
47
 
48
  ### Training results
49
 
50
+ | Training Loss | Epoch | Step | Validation Loss |
51
+ |:-------------:|:------:|:----:|:---------------:|
52
+ | 1.2197 | 0.7477 | 500 | 1.1066 |
53
 
54
 
55
  ### Framework versions
56
 
57
+ - Transformers 4.40.1
58
  - Pytorch 2.2.1+cu121
59
+ - Datasets 2.19.0
60
+ - Tokenizers 0.19.1
config.json CHANGED
@@ -54,7 +54,7 @@
54
  "pad_token_id": 1,
55
  "repetition_penalty": 3.5,
56
  "torch_dtype": "float32",
57
- "transformers_version": "4.38.2",
58
  "use_cache": true,
59
  "vocab_size": 50265
60
  }
 
54
  "pad_token_id": 1,
55
  "repetition_penalty": 3.5,
56
  "torch_dtype": "float32",
57
+ "transformers_version": "4.40.1",
58
  "use_cache": true,
59
  "vocab_size": 50265
60
  }
generation_config.json CHANGED
@@ -11,5 +11,5 @@
11
  "num_beams": 4,
12
  "pad_token_id": 1,
13
  "repetition_penalty": 3.5,
14
- "transformers_version": "4.38.2"
15
  }
 
11
  "num_beams": 4,
12
  "pad_token_id": 1,
13
  "repetition_penalty": 3.5,
14
+ "transformers_version": "4.40.1"
15
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16a5d2233c5752932dfdcd6f88d2fd1a3f6dad79c7083fcd197f968414e459b1
3
  size 647614116
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61917097c9cd1d775a9cc1a28ea6478e5038b4731383b1caa0289c3b4769142c
3
  size 647614116
tokenizer.json CHANGED
@@ -97,6 +97,7 @@
97
  "end_of_word_suffix": "",
98
  "fuse_unk": false,
99
  "byte_fallback": false,
 
100
  "vocab": {
101
  "<s>": 0,
102
  "<pad>": 1,
 
97
  "end_of_word_suffix": "",
98
  "fuse_unk": false,
99
  "byte_fallback": false,
100
+ "ignore_merges": false,
101
  "vocab": {
102
  "<s>": 0,
103
  "<pad>": 1,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:37954f477c05d31021d47ec35c1266fc46e0573105c6226de46ed96e5f543cb7
3
- size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07a5e6df1dcd5b8c6dc862e06712e4ef081606fb6b89139367d99fa7a57170ed
3
+ size 4984