SumanNazir commited on
Commit
5817cd1
1 Parent(s): 1d007e8

Training in progress, step 500

Browse files
README.md CHANGED
@@ -1,11 +1,9 @@
1
  ---
2
  library_name: transformers
3
  license: mit
4
- base_model: SumanNazir/highlight_summary_model_trained_on_reduced_data
5
  tags:
6
  - generated_from_trainer
7
- metrics:
8
- - rouge
9
  model-index:
10
  - name: highlight_summary_model_trained_on_reduced_data
11
  results: []
@@ -16,14 +14,9 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # highlight_summary_model_trained_on_reduced_data
18
 
19
- This model is a fine-tuned version of [SumanNazir/highlight_summary_model_trained_on_reduced_data](https://huggingface.co/SumanNazir/highlight_summary_model_trained_on_reduced_data) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 4.7393
22
- - Rouge1: 0.4167
23
- - Rouge2: 0.1684
24
- - Rougel: 0.3095
25
- - Rougelsum: 0.3088
26
- - Generated Length: 38.5758
27
 
28
  ## Model description
29
 
@@ -43,8 +36,8 @@ More information needed
43
 
44
  The following hyperparameters were used during training:
45
  - learning_rate: 2e-05
46
- - train_batch_size: 2
47
- - eval_batch_size: 2
48
  - seed: 42
49
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
  - lr_scheduler_type: linear
@@ -53,15 +46,15 @@ The following hyperparameters were used during training:
53
 
54
  ### Training results
55
 
56
- | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Generated Length |
57
- |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:----------------:|
58
- | No log | 1.0 | 263 | 4.6745 | 0.4127 | 0.1629 | 0.3115 | 0.3119 | 38.0909 |
59
- | 0.0309 | 2.0 | 526 | 4.7393 | 0.4167 | 0.1684 | 0.3095 | 0.3088 | 38.5758 |
60
 
61
 
62
  ### Framework versions
63
 
64
- - Transformers 4.45.1
65
  - Pytorch 2.4.1+cu121
66
  - Datasets 3.0.1
67
- - Tokenizers 0.20.0
 
1
  ---
2
  library_name: transformers
3
  license: mit
4
+ base_model: facebook/bart-large-xsum
5
  tags:
6
  - generated_from_trainer
 
 
7
  model-index:
8
  - name: highlight_summary_model_trained_on_reduced_data
9
  results: []
 
14
 
15
  # highlight_summary_model_trained_on_reduced_data
16
 
17
+ This model is a fine-tuned version of [facebook/bart-large-xsum](https://huggingface.co/facebook/bart-large-xsum) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 1.7400
 
 
 
 
 
20
 
21
  ## Model description
22
 
 
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 2e-05
39
+ - train_batch_size: 8
40
+ - eval_batch_size: 8
41
  - seed: 42
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
 
46
 
47
  ### Training results
48
 
49
+ | Training Loss | Epoch | Step | Validation Loss |
50
+ |:-------------:|:-----:|:----:|:---------------:|
51
+ | No log | 1.0 | 274 | 1.7445 |
52
+ | 1.6984 | 2.0 | 548 | 1.7400 |
53
 
54
 
55
  ### Framework versions
56
 
57
+ - Transformers 4.44.2
58
  - Pytorch 2.4.1+cu121
59
  - Datasets 3.0.1
60
+ - Tokenizers 0.19.1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "SumanNazir/highlight_summary_model_trained_on_reduced_data",
3
  "_num_labels": 3,
4
  "activation_dropout": 0.0,
5
  "activation_function": "gelu",
@@ -19,7 +19,7 @@
19
  "decoder_layers": 12,
20
  "decoder_start_token_id": 2,
21
  "dropout": 0.1,
22
- "early_stopping": null,
23
  "encoder_attention_heads": 16,
24
  "encoder_ffn_dim": 4096,
25
  "encoder_layerdrop": 0.0,
@@ -42,14 +42,14 @@
42
  "LABEL_1": 1,
43
  "LABEL_2": 2
44
  },
45
- "max_length": null,
46
  "max_position_embeddings": 1024,
47
- "min_length": null,
48
  "model_type": "bart",
49
- "no_repeat_ngram_size": null,
50
  "normalize_before": false,
51
  "normalize_embedding": true,
52
- "num_beams": null,
53
  "num_hidden_layers": 12,
54
  "output_past": true,
55
  "pad_token_id": 1,
@@ -61,7 +61,7 @@
61
  "student_encoder_layers": null,
62
  "task_specific_params": {},
63
  "torch_dtype": "float32",
64
- "transformers_version": "4.45.1",
65
  "use_cache": true,
66
  "vocab_size": 50264
67
  }
 
1
  {
2
+ "_name_or_path": "facebook/bart-large-xsum",
3
  "_num_labels": 3,
4
  "activation_dropout": 0.0,
5
  "activation_function": "gelu",
 
19
  "decoder_layers": 12,
20
  "decoder_start_token_id": 2,
21
  "dropout": 0.1,
22
+ "early_stopping": true,
23
  "encoder_attention_heads": 16,
24
  "encoder_ffn_dim": 4096,
25
  "encoder_layerdrop": 0.0,
 
42
  "LABEL_1": 1,
43
  "LABEL_2": 2
44
  },
45
+ "max_length": 62,
46
  "max_position_embeddings": 1024,
47
+ "min_length": 11,
48
  "model_type": "bart",
49
+ "no_repeat_ngram_size": 3,
50
  "normalize_before": false,
51
  "normalize_embedding": true,
52
+ "num_beams": 6,
53
  "num_hidden_layers": 12,
54
  "output_past": true,
55
  "pad_token_id": 1,
 
61
  "student_encoder_layers": null,
62
  "task_specific_params": {},
63
  "torch_dtype": "float32",
64
+ "transformers_version": "4.44.2",
65
  "use_cache": true,
66
  "vocab_size": 50264
67
  }
generation_config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "bos_token_id": 0,
3
  "decoder_start_token_id": 2,
4
  "early_stopping": true,
@@ -9,5 +10,5 @@
9
  "no_repeat_ngram_size": 3,
10
  "num_beams": 6,
11
  "pad_token_id": 1,
12
- "transformers_version": "4.45.1"
13
  }
 
1
  {
2
+ "_from_model_config": true,
3
  "bos_token_id": 0,
4
  "decoder_start_token_id": 2,
5
  "early_stopping": true,
 
10
  "no_repeat_ngram_size": 3,
11
  "num_beams": 6,
12
  "pad_token_id": 1,
13
+ "transformers_version": "4.44.2"
14
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dae807a36ab2a73745aa6a8774bf07f05777b341764cb7cf949a7ec82109c834
3
  size 1625422896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec831245c1c44fa9b8c40b10528656362ca2657d808f466ae3e07000a3b39745
3
  size 1625422896
runs/Sep30_20-26-55_13959885f477/events.out.tfevents.1727728016.13959885f477.218.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65fb1689147f1aab9cb9d31dc3605174ac98acf95cb2a8541048e02ceba39e10
3
+ size 6300
special_tokens_map.json CHANGED
@@ -1,25 +1,7 @@
1
  {
2
- "bos_token": {
3
- "content": "<s>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "cls_token": {
10
- "content": "<s>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "eos_token": {
17
- "content": "</s>",
18
- "lstrip": false,
19
- "normalized": true,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
  "mask_token": {
24
  "content": "<mask>",
25
  "lstrip": true,
@@ -27,25 +9,7 @@
27
  "rstrip": false,
28
  "single_word": false
29
  },
30
- "pad_token": {
31
- "content": "<pad>",
32
- "lstrip": false,
33
- "normalized": true,
34
- "rstrip": false,
35
- "single_word": false
36
- },
37
- "sep_token": {
38
- "content": "</s>",
39
- "lstrip": false,
40
- "normalized": true,
41
- "rstrip": false,
42
- "single_word": false
43
- },
44
- "unk_token": {
45
- "content": "<unk>",
46
- "lstrip": false,
47
- "normalized": true,
48
- "rstrip": false,
49
- "single_word": false
50
- }
51
  }
 
1
  {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "mask_token": {
6
  "content": "<mask>",
7
  "lstrip": true,
 
9
  "rstrip": false,
10
  "single_word": false
11
  },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -48,14 +48,10 @@
48
  "eos_token": "</s>",
49
  "errors": "replace",
50
  "mask_token": "<mask>",
51
- "max_length": 128,
52
  "model_max_length": 1024,
53
  "pad_token": "<pad>",
54
  "sep_token": "</s>",
55
- "stride": 0,
56
  "tokenizer_class": "BartTokenizer",
57
  "trim_offsets": true,
58
- "truncation_side": "right",
59
- "truncation_strategy": "longest_first",
60
  "unk_token": "<unk>"
61
  }
 
48
  "eos_token": "</s>",
49
  "errors": "replace",
50
  "mask_token": "<mask>",
 
51
  "model_max_length": 1024,
52
  "pad_token": "<pad>",
53
  "sep_token": "</s>",
 
54
  "tokenizer_class": "BartTokenizer",
55
  "trim_offsets": true,
 
 
56
  "unk_token": "<unk>"
57
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:040a829b06db162606556c2f6239102a308d4530eaa19d03f1bf19f978ba81dd
3
- size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6582dc6f594bf01c19035f3963975b4f785d0bc02bdc2f50ec12495118b5d2af
3
+ size 5368