scottsuk0306 commited on
Commit
647efee
1 Parent(s): 61d6e02

Model save

Browse files
README.md CHANGED
@@ -2,7 +2,6 @@
2
  license: gemma
3
  base_model: scottsuk0306/easylm-sft-gemma-2-2b
4
  tags:
5
- - easylm
6
  - trl
7
  - reward-trainer
8
  - generated_from_trainer
@@ -16,13 +15,12 @@ model-index:
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
  should probably proofread and complete it, then remove this comment. -->
18
 
19
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/lklab_kaist/grm/runs/kp392ohf)
20
  # easylm-rm-gemma-2-2b
21
 
22
  This model is a fine-tuned version of [scottsuk0306/easylm-sft-gemma-2-2b](https://huggingface.co/scottsuk0306/easylm-sft-gemma-2-2b) on an unknown dataset.
23
  It achieves the following results on the evaluation set:
24
- - Loss: 1.7170
25
- - Accuracy: 0.6042
26
 
27
  ## Model description
28
 
@@ -57,14 +55,14 @@ The following hyperparameters were used during training:
57
 
58
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
59
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
60
- | No log | 1.0 | 25 | 0.6833 | 0.5417 |
61
- | No log | 2.0 | 50 | 1.2481 | 0.5625 |
62
- | No log | 3.0 | 75 | 1.7170 | 0.6042 |
63
 
64
 
65
  ### Framework versions
66
 
67
- - Transformers 4.42.4
68
  - Pytorch 2.3.0+cu121
69
  - Datasets 2.20.0
70
  - Tokenizers 0.19.1
 
2
  license: gemma
3
  base_model: scottsuk0306/easylm-sft-gemma-2-2b
4
  tags:
 
5
  - trl
6
  - reward-trainer
7
  - generated_from_trainer
 
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
  should probably proofread and complete it, then remove this comment. -->
17
 
 
18
  # easylm-rm-gemma-2-2b
19
 
20
  This model is a fine-tuned version of [scottsuk0306/easylm-sft-gemma-2-2b](https://huggingface.co/scottsuk0306/easylm-sft-gemma-2-2b) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 6.0469
23
+ - Accuracy: 0.5833
24
 
25
  ## Model description
26
 
 
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
57
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
58
+ | 0.6795 | 1.0 | 2399 | 0.7247 | 0.5938 |
59
+ | 0.4438 | 2.0 | 4798 | 2.9242 | 0.5938 |
60
+ | 0.0414 | 3.0 | 7197 | 6.0469 | 0.5833 |
61
 
62
 
63
  ### Framework versions
64
 
65
+ - Transformers 4.43.3
66
  - Pytorch 2.3.0+cu121
67
  - Datasets 2.20.0
68
  - Tokenizers 0.19.1
all_results.json CHANGED
@@ -1,15 +1,9 @@
1
  {
2
  "epoch": 3.0,
3
- "eval_accuracy": 0.6041666666666666,
4
- "eval_loss": 1.7170292139053345,
5
- "eval_runtime": 1.6843,
6
- "eval_samples": 96,
7
- "eval_samples_per_second": 56.996,
8
- "eval_steps_per_second": 14.249,
9
  "total_flos": 0.0,
10
- "train_loss": 0.5482939656575521,
11
- "train_runtime": 190.4657,
12
- "train_samples": 100,
13
- "train_samples_per_second": 1.575,
14
- "train_steps_per_second": 0.394
15
  }
 
1
  {
2
  "epoch": 3.0,
 
 
 
 
 
 
3
  "total_flos": 0.0,
4
+ "train_loss": 0.38497807171802156,
5
+ "train_runtime": 4417.8859,
6
+ "train_samples": 9595,
7
+ "train_samples_per_second": 6.516,
8
+ "train_steps_per_second": 1.629
9
  }
config.json CHANGED
@@ -33,7 +33,7 @@
33
  "rope_theta": 10000.0,
34
  "sliding_window": 4096,
35
  "torch_dtype": "bfloat16",
36
- "transformers_version": "4.42.4",
37
  "use_cache": true,
38
  "vocab_size": 256000
39
  }
 
33
  "rope_theta": 10000.0,
34
  "sliding_window": 4096,
35
  "torch_dtype": "bfloat16",
36
+ "transformers_version": "4.43.3",
37
  "use_cache": true,
38
  "vocab_size": 256000
39
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b00ac57221c6afcf0c9d6b84dd92b855c0fde0c5bef251a4bf68be9462cf5f8a
3
  size 4988025760
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebdc412ae8e6303311b7295afc2426709718573671e7dc3fbbb856569273e82b
3
  size 4988025760
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f7dc4b3fadae1742a61cbcb9c8f61144918b5b86476fb7e673979b4670ee6502
3
  size 240696424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab88370ef0a7abfcf32f7b00c3788ae7e30f00749c2146b0fafdb348c75eecc2
3
  size 240696424
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 3.0,
3
  "total_flos": 0.0,
4
- "train_loss": 0.5482939656575521,
5
- "train_runtime": 190.4657,
6
- "train_samples": 100,
7
- "train_samples_per_second": 1.575,
8
- "train_steps_per_second": 0.394
9
  }
 
1
  {
2
  "epoch": 3.0,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.38497807171802156,
5
+ "train_runtime": 4417.8859,
6
+ "train_samples": 9595,
7
+ "train_samples_per_second": 6.516,
8
+ "train_steps_per_second": 1.629
9
  }
trainer_state.json CHANGED
@@ -3,50 +3,148 @@
3
  "best_model_checkpoint": null,
4
  "epoch": 3.0,
5
  "eval_steps": 500,
6
- "global_step": 75,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  {
12
  "epoch": 1.0,
13
- "eval_accuracy": 0.5416666666666666,
14
- "eval_loss": 0.6833394169807434,
15
- "eval_runtime": 1.6468,
16
- "eval_samples_per_second": 58.294,
17
- "eval_steps_per_second": 14.574,
18
- "step": 25
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  },
20
  {
21
  "epoch": 2.0,
22
- "eval_accuracy": 0.5625,
23
- "eval_loss": 1.2481266260147095,
24
- "eval_runtime": 1.672,
25
- "eval_samples_per_second": 57.415,
26
- "eval_steps_per_second": 14.354,
27
- "step": 50
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
  },
29
  {
30
  "epoch": 3.0,
31
- "eval_accuracy": 0.6041666666666666,
32
- "eval_loss": 1.7170292139053345,
33
- "eval_runtime": 1.7763,
34
- "eval_samples_per_second": 54.045,
35
- "eval_steps_per_second": 13.511,
36
- "step": 75
37
  },
38
  {
39
  "epoch": 3.0,
40
- "step": 75,
41
  "total_flos": 0.0,
42
- "train_loss": 0.5482939656575521,
43
- "train_runtime": 190.4657,
44
- "train_samples_per_second": 1.575,
45
- "train_steps_per_second": 0.394
46
  }
47
  ],
48
  "logging_steps": 500,
49
- "max_steps": 75,
50
  "num_input_tokens_seen": 0,
51
  "num_train_epochs": 3,
52
  "save_steps": 500,
@@ -56,8 +154,8 @@
56
  "should_epoch_stop": false,
57
  "should_evaluate": false,
58
  "should_log": false,
59
- "should_save": true,
60
- "should_training_stop": true
61
  },
62
  "attributes": {}
63
  }
 
3
  "best_model_checkpoint": null,
4
  "epoch": 3.0,
5
  "eval_steps": 500,
6
+ "global_step": 7197,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
+ {
12
+ "epoch": 0.20842017507294705,
13
+ "grad_norm": 7.509331703186035,
14
+ "learning_rate": 2.9644144799482153e-06,
15
+ "loss": 0.7098,
16
+ "step": 500
17
+ },
18
+ {
19
+ "epoch": 0.4168403501458941,
20
+ "grad_norm": 5.73619270324707,
21
+ "learning_rate": 2.859346358776001e-06,
22
+ "loss": 0.6948,
23
+ "step": 1000
24
+ },
25
+ {
26
+ "epoch": 0.6252605252188412,
27
+ "grad_norm": 7.396816253662109,
28
+ "learning_rate": 2.689780841460394e-06,
29
+ "loss": 0.6855,
30
+ "step": 1500
31
+ },
32
+ {
33
+ "epoch": 0.8336807002917882,
34
+ "grad_norm": 9.113743782043457,
35
+ "learning_rate": 2.4637633641567626e-06,
36
+ "loss": 0.6795,
37
+ "step": 2000
38
+ },
39
  {
40
  "epoch": 1.0,
41
+ "eval_accuracy": 0.59375,
42
+ "eval_loss": 0.7246907353401184,
43
+ "eval_runtime": 1.6903,
44
+ "eval_samples_per_second": 56.793,
45
+ "eval_steps_per_second": 14.198,
46
+ "step": 2399
47
+ },
48
+ {
49
+ "epoch": 1.0421008753647354,
50
+ "grad_norm": 33.38862609863281,
51
+ "learning_rate": 2.1920178594926297e-06,
52
+ "loss": 0.6189,
53
+ "step": 2500
54
+ },
55
+ {
56
+ "epoch": 1.2505210504376825,
57
+ "grad_norm": 3.111920118331909,
58
+ "learning_rate": 1.8874379342749385e-06,
59
+ "loss": 0.4549,
60
+ "step": 3000
61
+ },
62
+ {
63
+ "epoch": 1.4589412255106295,
64
+ "grad_norm": 0.5629640221595764,
65
+ "learning_rate": 1.5644751018852966e-06,
66
+ "loss": 0.4627,
67
+ "step": 3500
68
+ },
69
+ {
70
+ "epoch": 1.6673614005835766,
71
+ "grad_norm": 0.07236230373382568,
72
+ "learning_rate": 1.2384530961210145e-06,
73
+ "loss": 0.4569,
74
+ "step": 4000
75
+ },
76
+ {
77
+ "epoch": 1.8757815756565237,
78
+ "grad_norm": 0.02289917692542076,
79
+ "learning_rate": 9.24840800480023e-07,
80
+ "loss": 0.4438,
81
+ "step": 4500
82
  },
83
  {
84
  "epoch": 2.0,
85
+ "eval_accuracy": 0.59375,
86
+ "eval_loss": 2.9241831302642822,
87
+ "eval_runtime": 1.6517,
88
+ "eval_samples_per_second": 58.123,
89
+ "eval_steps_per_second": 14.531,
90
+ "step": 4798
91
+ },
92
+ {
93
+ "epoch": 2.084201750729471,
94
+ "grad_norm": 4.127718925476074,
95
+ "learning_rate": 6.385182904756809e-07,
96
+ "loss": 0.2538,
97
+ "step": 5000
98
+ },
99
+ {
100
+ "epoch": 2.292621925802418,
101
+ "grad_norm": 1.510969994955591e-11,
102
+ "learning_rate": 3.93070813336037e-07,
103
+ "loss": 0.0133,
104
+ "step": 5500
105
+ },
106
+ {
107
+ "epoch": 2.501042100875365,
108
+ "grad_norm": 3.9616837588241083e-10,
109
+ "learning_rate": 2.001442038869754e-07,
110
+ "loss": 0.0213,
111
+ "step": 6000
112
+ },
113
+ {
114
+ "epoch": 2.7094622759483116,
115
+ "grad_norm": 1.3078284837320098e-06,
116
+ "learning_rate": 6.889232043392568e-08,
117
+ "loss": 0.0046,
118
+ "step": 6500
119
+ },
120
+ {
121
+ "epoch": 2.917882451021259,
122
+ "grad_norm": 3.3587113534849777e-07,
123
+ "learning_rate": 5.5427183508257595e-09,
124
+ "loss": 0.0414,
125
+ "step": 7000
126
  },
127
  {
128
  "epoch": 3.0,
129
+ "eval_accuracy": 0.5833333333333334,
130
+ "eval_loss": 6.046913146972656,
131
+ "eval_runtime": 1.6474,
132
+ "eval_samples_per_second": 58.275,
133
+ "eval_steps_per_second": 14.569,
134
+ "step": 7197
135
  },
136
  {
137
  "epoch": 3.0,
138
+ "step": 7197,
139
  "total_flos": 0.0,
140
+ "train_loss": 0.38497807171802156,
141
+ "train_runtime": 4417.8859,
142
+ "train_samples_per_second": 6.516,
143
+ "train_steps_per_second": 1.629
144
  }
145
  ],
146
  "logging_steps": 500,
147
+ "max_steps": 7197,
148
  "num_input_tokens_seen": 0,
149
  "num_train_epochs": 3,
150
  "save_steps": 500,
 
154
  "should_epoch_stop": false,
155
  "should_evaluate": false,
156
  "should_log": false,
157
+ "should_save": false,
158
+ "should_training_stop": false
159
  },
160
  "attributes": {}
161
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:17509e8d66be67771ace016820380128ce14fad5edf7c3939a29014c02226a57
3
- size 6200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3a81758449ae164f1bdd71e177b28bb7eae16b05cdabb93ec9211a34cf70b14
3
+ size 6264