model
Browse files- scripts/model.yaml +5 -3
scripts/model.yaml
CHANGED
@@ -38,7 +38,8 @@ initial_checkpoint_dir:
|
|
38 |
# from the latest checkpoint in ``out_dir``. An error will be raised if no checkpoint is found. Passing
|
39 |
# ``'auto'`` will resume from the latest checkpoint but not error if no checkpoint exists.
|
40 |
# (type: Union[bool, Literal["auto"], Path], default: False)
|
41 |
-
resume: false
|
|
|
42 |
|
43 |
# Data-related arguments. If not provided, the default is ``litgpt.data.TinyLlama``.
|
44 |
# data: LitData
|
@@ -61,10 +62,11 @@ train:
|
|
61 |
lr_warmup_steps: 2000
|
62 |
|
63 |
# Number of epochs to train on (type: Optional[int], default: null)
|
64 |
-
epochs:
|
65 |
|
66 |
# Total number of tokens to train on (type: Optional[int], default: 3000000000000)
|
67 |
-
max_tokens: 3000000000000
|
|
|
68 |
|
69 |
# Limits the number of optimizer steps to run. (type: Optional[int], default: null)
|
70 |
max_steps:
|
|
|
38 |
# from the latest checkpoint in ``out_dir``. An error will be raised if no checkpoint is found. Passing
|
39 |
# ``'auto'`` will resume from the latest checkpoint but not error if no checkpoint exists.
|
40 |
# (type: Union[bool, Literal["auto"], Path], default: False)
|
41 |
+
# resume: false
|
42 |
+
resume: "auto"
|
43 |
|
44 |
# Data-related arguments. If not provided, the default is ``litgpt.data.TinyLlama``.
|
45 |
# data: LitData
|
|
|
62 |
lr_warmup_steps: 2000
|
63 |
|
64 |
# Number of epochs to train on (type: Optional[int], default: null)
|
65 |
+
epochs: 3
|
66 |
|
67 |
# Total number of tokens to train on (type: Optional[int], default: 3000000000000)
|
68 |
+
# max_tokens: 3000000000000
|
69 |
+
max_tokens: 3260735571 # 1591379 * 2049
|
70 |
|
71 |
# Limits the number of optimizer steps to run. (type: Optional[int], default: null)
|
72 |
max_steps:
|