bitsoko commited on
Commit
8b1c692
·
verified ·
1 Parent(s): 577a365

Training in progress, step 50, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -20,12 +20,12 @@
20
  "rank_pattern": {},
21
  "revision": "unsloth",
22
  "target_modules": [
23
- "k_proj",
 
24
  "v_proj",
25
  "up_proj",
26
- "down_proj",
27
  "q_proj",
28
- "o_proj",
29
  "gate_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
 
20
  "rank_pattern": {},
21
  "revision": "unsloth",
22
  "target_modules": [
23
+ "down_proj",
24
+ "o_proj",
25
  "v_proj",
26
  "up_proj",
27
+ "k_proj",
28
  "q_proj",
 
29
  "gate_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:53555ee8a4eb110c4bdce3fd50ed70e1c37d2228502979597728308e7dbc4c8a
3
  size 1912664024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f79c51cd876c8b5b6f891bdff8fd168dd531ab20717f8389982574d19f0dbea0
3
  size 1912664024
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1645b3ce8c7b797e757add674024ac2f3c2bdadc80e8a598a873bcc34e1ba413
3
  size 958697364
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bf021cd132f16efcbc21f9c136335d5522103e91e16b4f5562176a75367b8c4
3
  size 958697364
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:caf99753320d5a922d21785c9acff8aad90fb47892dad1f221c94717bf3a04b1
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:376fa8cdca401e75879a433c4cf57e5c31b5973d7635f7e690ffeb05b513101b
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:21226719b9e25fe4304ec632d533e056eb568a7272feb0ef7163c162da3912fa
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4039ca40b296c95519ba3582af20e2ac85fa629c760519260b975e4ffc9aa9b
3
  size 1064
last-checkpoint/tokenizer_config.json CHANGED
@@ -122,7 +122,7 @@
122
  "legacy": false,
123
  "model_max_length": 4096,
124
  "pad_token": "<|placeholder6|>",
125
- "padding_side": "right",
126
  "sp_model_kwargs": {},
127
  "tokenizer_class": "LlamaTokenizer",
128
  "unk_token": "<unk>",
 
122
  "legacy": false,
123
  "model_max_length": 4096,
124
  "pad_token": "<|placeholder6|>",
125
+ "padding_side": "left",
126
  "sp_model_kwargs": {},
127
  "tokenizer_class": "LlamaTokenizer",
128
  "unk_token": "<unk>",
last-checkpoint/trainer_state.json CHANGED
@@ -1,170 +1,50 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.012270691453463402,
5
  "eval_steps": 20,
6
- "global_step": 200,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.0012270691453463403,
13
- "grad_norm": 0.05003603920340538,
14
- "learning_rate": 0.00019993864026834657,
15
- "loss": 2.6971,
16
  "step": 20
17
  },
18
  {
19
  "epoch": 0.0012270691453463403,
20
- "eval_loss": 2.296663284301758,
21
- "eval_runtime": 26.0065,
22
- "eval_samples_per_second": 3.845,
23
- "eval_steps_per_second": 0.5,
24
  "step": 20
25
  },
26
  {
27
  "epoch": 0.0024541382906926807,
28
- "grad_norm": 0.07180392742156982,
29
- "learning_rate": 0.00019985682729280866,
30
- "loss": 2.202,
31
  "step": 40
32
  },
33
  {
34
  "epoch": 0.0024541382906926807,
35
- "eval_loss": 2.067845582962036,
36
- "eval_runtime": 26.2005,
37
- "eval_samples_per_second": 3.817,
38
- "eval_steps_per_second": 0.496,
39
  "step": 40
40
- },
41
- {
42
- "epoch": 0.003681207436039021,
43
- "grad_norm": 0.08042994141578674,
44
- "learning_rate": 0.00019977501431727072,
45
- "loss": 2.0527,
46
- "step": 60
47
- },
48
- {
49
- "epoch": 0.003681207436039021,
50
- "eval_loss": 1.9334843158721924,
51
- "eval_runtime": 25.7281,
52
- "eval_samples_per_second": 3.887,
53
- "eval_steps_per_second": 0.505,
54
- "step": 60
55
- },
56
- {
57
- "epoch": 0.004908276581385361,
58
- "grad_norm": 0.08663463592529297,
59
- "learning_rate": 0.0001996932013417328,
60
- "loss": 1.9392,
61
- "step": 80
62
- },
63
- {
64
- "epoch": 0.004908276581385361,
65
- "eval_loss": 1.8688364028930664,
66
- "eval_runtime": 26.145,
67
- "eval_samples_per_second": 3.825,
68
- "eval_steps_per_second": 0.497,
69
- "step": 80
70
- },
71
- {
72
- "epoch": 0.006135345726731701,
73
- "grad_norm": 0.08462908118963242,
74
- "learning_rate": 0.00019961138836619487,
75
- "loss": 1.8772,
76
- "step": 100
77
- },
78
- {
79
- "epoch": 0.006135345726731701,
80
- "eval_loss": 1.8178932666778564,
81
- "eval_runtime": 26.3004,
82
- "eval_samples_per_second": 3.802,
83
- "eval_steps_per_second": 0.494,
84
- "step": 100
85
- },
86
- {
87
- "epoch": 0.007362414872078042,
88
- "grad_norm": 0.11603377014398575,
89
- "learning_rate": 0.000199529575390657,
90
- "loss": 1.7827,
91
- "step": 120
92
- },
93
- {
94
- "epoch": 0.007362414872078042,
95
- "eval_loss": 1.773740291595459,
96
- "eval_runtime": 26.4277,
97
- "eval_samples_per_second": 3.784,
98
- "eval_steps_per_second": 0.492,
99
- "step": 120
100
- },
101
- {
102
- "epoch": 0.008589484017424381,
103
- "grad_norm": 0.10753527283668518,
104
- "learning_rate": 0.00019944776241511905,
105
- "loss": 1.8118,
106
- "step": 140
107
- },
108
- {
109
- "epoch": 0.008589484017424381,
110
- "eval_loss": 1.7376396656036377,
111
- "eval_runtime": 26.2097,
112
- "eval_samples_per_second": 3.815,
113
- "eval_steps_per_second": 0.496,
114
- "step": 140
115
- },
116
- {
117
- "epoch": 0.009816553162770723,
118
- "grad_norm": 0.09732338786125183,
119
- "learning_rate": 0.00019936594943958114,
120
- "loss": 1.7397,
121
- "step": 160
122
- },
123
- {
124
- "epoch": 0.009816553162770723,
125
- "eval_loss": 1.7010047435760498,
126
- "eval_runtime": 25.8266,
127
- "eval_samples_per_second": 3.872,
128
- "eval_steps_per_second": 0.503,
129
- "step": 160
130
- },
131
- {
132
- "epoch": 0.011043622308117063,
133
- "grad_norm": 0.11193066835403442,
134
- "learning_rate": 0.0001992841364640432,
135
- "loss": 1.7868,
136
- "step": 180
137
- },
138
- {
139
- "epoch": 0.011043622308117063,
140
- "eval_loss": 1.6764620542526245,
141
- "eval_runtime": 26.1757,
142
- "eval_samples_per_second": 3.82,
143
- "eval_steps_per_second": 0.497,
144
- "step": 180
145
- },
146
- {
147
- "epoch": 0.012270691453463402,
148
- "grad_norm": 0.10227545350790024,
149
- "learning_rate": 0.0001992023234885053,
150
- "loss": 1.7799,
151
- "step": 200
152
- },
153
- {
154
- "epoch": 0.012270691453463402,
155
- "eval_loss": 1.656005859375,
156
- "eval_runtime": 26.2305,
157
- "eval_samples_per_second": 3.812,
158
- "eval_steps_per_second": 0.496,
159
- "step": 200
160
  }
161
  ],
162
  "logging_steps": 20,
163
- "max_steps": 48897,
164
  "num_input_tokens_seen": 0,
165
- "num_train_epochs": 3,
166
  "save_steps": 50,
167
- "total_flos": 1.41326994253824e+16,
168
  "train_batch_size": 2,
169
  "trial_name": null,
170
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.0030676728633658506,
5
  "eval_steps": 20,
6
+ "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.0012270691453463403,
13
+ "grad_norm": NaN,
14
+ "learning_rate": 0.0,
15
+ "loss": 3.3245,
16
  "step": 20
17
  },
18
  {
19
  "epoch": 0.0012270691453463403,
20
+ "eval_loss": 1.7450172901153564,
21
+ "eval_runtime": 24.6271,
22
+ "eval_samples_per_second": 4.061,
23
+ "eval_steps_per_second": 0.528,
24
  "step": 20
25
  },
26
  {
27
  "epoch": 0.0024541382906926807,
28
+ "grad_norm": 0.0,
29
+ "learning_rate": 0.0001999263532588683,
30
+ "loss": 3.5708,
31
  "step": 40
32
  },
33
  {
34
  "epoch": 0.0024541382906926807,
35
+ "eval_loss": 1.7450109720230103,
36
+ "eval_runtime": 26.043,
37
+ "eval_samples_per_second": 3.84,
38
+ "eval_steps_per_second": 0.499,
39
  "step": 40
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  }
41
  ],
42
  "logging_steps": 20,
43
+ "max_steps": 16299,
44
  "num_input_tokens_seen": 0,
45
+ "num_train_epochs": 1,
46
  "save_steps": 50,
47
+ "total_flos": 3253280243097600.0,
48
  "train_batch_size": 2,
49
  "trial_name": null,
50
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5e05a5a439e5cc5d47d8ef9a0303b090d4a155ad12a03f09a5d603bb3e1e1d88
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:564d5c4c71e774ca6ed6b6056431bf55581adabc7617b07cefd97a1c116e6504
3
  size 4920