ramdhanfirdaus commited on
Commit
edcf2ba
1 Parent(s): 5d93b1d

Training in progress, step 1800, checkpoint

Browse files
last-checkpoint/README.md CHANGED
@@ -201,18 +201,6 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
201
  ## Training procedure
202
 
203
 
204
- The following `bitsandbytes` quantization config was used during training:
205
- - quant_method: bitsandbytes
206
- - load_in_8bit: False
207
- - load_in_4bit: True
208
- - llm_int8_threshold: 6.0
209
- - llm_int8_skip_modules: None
210
- - llm_int8_enable_fp32_cpu_offload: False
211
- - llm_int8_has_fp16_weight: False
212
- - bnb_4bit_quant_type: nf4
213
- - bnb_4bit_use_double_quant: True
214
- - bnb_4bit_compute_dtype: float16
215
-
216
  ### Framework versions
217
 
218
 
 
201
  ## Training procedure
202
 
203
 
 
 
 
 
 
 
 
 
 
 
 
 
204
  ### Framework versions
205
 
206
 
last-checkpoint/adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c3694b9769451b4e7587f8afd0bf460e687c5c552683b2b6a8832e5d0943cd6
3
  size 50349441
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0e617333294c663e0e8e0a9c6128eec6573f75784de9b79819e9c3c76627aae
3
  size 50349441
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f211b36881b86340c0e1c1926f46a2a99a1d45eb54e2a242690a4e07cddb2cf6
3
- size 100693001
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62ce02d027a3d5884b9e99ddaa4ac0ac1441dd8e99dca2d29a8692a2980e52bb
3
+ size 100691721
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1389c8fb4da5b08654d5ecff1c857ae43115866bad194f9289834fc2a0b6a9c8
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a85376be8f1e945d481585c1087e047e7112f253e5e6b615f10107365cf015a8
3
  size 14575
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0410ad60e6d0138b921a03b0a3e367fd27c6ab07cb9a5006fcb66ea8e5bbacc4
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0e0a1736a46fd1627af3c246e44261aaac909256abbd413b5ee5c968f6b2d8e
3
  size 627
last-checkpoint/special_tokens_map.json CHANGED
@@ -1,6 +1,24 @@
1
  {
2
- "bos_token": "<|endoftext|>",
3
- "eos_token": "<|endoftext|>",
 
 
 
 
 
 
 
 
 
 
 
 
4
  "pad_token": "<|endoftext|>",
5
- "unk_token": "<|endoftext|>"
 
 
 
 
 
 
6
  }
 
1
  {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
  "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
  }
last-checkpoint/tokenizer_config.json CHANGED
@@ -13,8 +13,12 @@
13
  "bos_token": "<|endoftext|>",
14
  "clean_up_tokenization_spaces": true,
15
  "eos_token": "<|endoftext|>",
 
16
  "model_max_length": 1024,
17
  "pad_token": "<|endoftext|>",
 
18
  "tokenizer_class": "GPT2Tokenizer",
 
 
19
  "unk_token": "<|endoftext|>"
20
  }
 
13
  "bos_token": "<|endoftext|>",
14
  "clean_up_tokenization_spaces": true,
15
  "eos_token": "<|endoftext|>",
16
+ "max_length": 512,
17
  "model_max_length": 1024,
18
  "pad_token": "<|endoftext|>",
19
+ "stride": 0,
20
  "tokenizer_class": "GPT2Tokenizer",
21
+ "truncation_side": "right",
22
+ "truncation_strategy": "longest_first",
23
  "unk_token": "<|endoftext|>"
24
  }
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 1.5012134313583374,
3
  "best_model_checkpoint": "./outputs/checkpoint-1700",
4
- "epoch": 1.238615664845173,
5
  "eval_steps": 100,
6
- "global_step": 1700,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -245,13 +245,27 @@
245
  "eval_samples_per_second": 43.564,
246
  "eval_steps_per_second": 5.451,
247
  "step": 1700
 
 
 
 
 
 
 
 
 
 
 
 
 
 
248
  }
249
  ],
250
  "logging_steps": 100,
251
  "max_steps": 4116,
252
  "num_train_epochs": 3,
253
  "save_steps": 100,
254
- "total_flos": 1.0135854358789325e+17,
255
  "trial_name": null,
256
  "trial_params": null
257
  }
 
1
  {
2
  "best_metric": 1.5012134313583374,
3
  "best_model_checkpoint": "./outputs/checkpoint-1700",
4
+ "epoch": 1.31183970856102,
5
  "eval_steps": 100,
6
+ "global_step": 1800,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
245
  "eval_samples_per_second": 43.564,
246
  "eval_steps_per_second": 5.451,
247
  "step": 1700
248
+ },
249
+ {
250
+ "epoch": 1.31,
251
+ "learning_rate": 0.0002,
252
+ "loss": 1.5144,
253
+ "step": 1800
254
+ },
255
+ {
256
+ "epoch": 1.31,
257
+ "eval_loss": 1.511275291442871,
258
+ "eval_runtime": 133.6708,
259
+ "eval_samples_per_second": 46.936,
260
+ "eval_steps_per_second": 5.873,
261
+ "step": 1800
262
  }
263
  ],
264
  "logging_steps": 100,
265
  "max_steps": 4116,
266
  "num_train_epochs": 3,
267
  "save_steps": 100,
268
+ "total_flos": 1.0731249717030912e+17,
269
  "trial_name": null,
270
  "trial_params": null
271
  }
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e42e5fbd3480211daf1554d169929e03bb21f025d78cef4fbd2e91363be51943
3
  size 4155
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6526a0e7745e6a6ee8517a3ecbb9cc450f19bb44a59493674b49cefdbde34f69
3
  size 4155