mtasic85 commited on
Commit
b347636
1 Parent(s): f711071
Files changed (1) hide show
  1. scripts/model.yaml +5 -3
scripts/model.yaml CHANGED
@@ -38,7 +38,8 @@ initial_checkpoint_dir:
38
  # from the latest checkpoint in ``out_dir``. An error will be raised if no checkpoint is found. Passing
39
  # ``'auto'`` will resume from the latest checkpoint but not error if no checkpoint exists.
40
  # (type: Union[bool, Literal["auto"], Path], default: False)
41
- resume: false
 
42
 
43
  # Data-related arguments. If not provided, the default is ``litgpt.data.TinyLlama``.
44
  # data: LitData
@@ -61,10 +62,11 @@ train:
61
  lr_warmup_steps: 2000
62
 
63
  # Number of epochs to train on (type: Optional[int], default: null)
64
- epochs:
65
 
66
  # Total number of tokens to train on (type: Optional[int], default: 3000000000000)
67
- max_tokens: 3000000000000
 
68
 
69
  # Limits the number of optimizer steps to run. (type: Optional[int], default: null)
70
  max_steps:
 
38
  # from the latest checkpoint in ``out_dir``. An error will be raised if no checkpoint is found. Passing
39
  # ``'auto'`` will resume from the latest checkpoint but not error if no checkpoint exists.
40
  # (type: Union[bool, Literal["auto"], Path], default: False)
41
+ # resume: false
42
+ resume: "auto"
43
 
44
  # Data-related arguments. If not provided, the default is ``litgpt.data.TinyLlama``.
45
  # data: LitData
 
62
  lr_warmup_steps: 2000
63
 
64
  # Number of epochs to train on (type: Optional[int], default: null)
65
+ epochs: 3
66
 
67
  # Total number of tokens to train on (type: Optional[int], default: 3000000000000)
68
+ # max_tokens: 3000000000000
69
+ max_tokens: 3260735571 # 1591379 * 2049
70
 
71
  # Limits the number of optimizer steps to run. (type: Optional[int], default: null)
72
  max_steps: