Spaces:
Paused
Paused
Update finetune.py
Browse files- finetune.py +13 -2
finetune.py
CHANGED
@@ -15,9 +15,19 @@ from peft import (
|
|
15 |
get_peft_model_state_dict,
|
16 |
)
|
17 |
|
18 |
-
|
19 |
-
# Parameters
|
20 |
HF_TOKEN = os.environ.get("TRL_TOKEN", None)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
21 |
MICRO_BATCH_SIZE = 16
|
22 |
BATCH_SIZE = 16
|
23 |
size = "7b"
|
@@ -153,6 +163,7 @@ trainer = transformers.Trainer(
|
|
153 |
save_total_limit=100,
|
154 |
load_best_model_at_end=True if VAL_SET_SIZE > 0 else False,
|
155 |
ddp_find_unused_parameters=False if ddp else None,
|
|
|
156 |
),
|
157 |
data_collator=transformers.DataCollatorForLanguageModeling(tokenizer, mlm=False),
|
158 |
)
|
|
|
15 |
get_peft_model_state_dict,
|
16 |
)
|
17 |
|
|
|
|
|
18 |
HF_TOKEN = os.environ.get("TRL_TOKEN", None)
|
19 |
+
if HF_TOKEN:
|
20 |
+
try:
|
21 |
+
shutil.rmtree("./data/")
|
22 |
+
except:
|
23 |
+
pass
|
24 |
+
|
25 |
+
repo = Repository(
|
26 |
+
local_dir="./data/", clone_from="gustavoaq/llama_ft", use_auth_token=HF_TOKEN, repo_type="models"
|
27 |
+
)
|
28 |
+
repo.git_pull()
|
29 |
+
# Parameters
|
30 |
+
|
31 |
MICRO_BATCH_SIZE = 16
|
32 |
BATCH_SIZE = 16
|
33 |
size = "7b"
|
|
|
163 |
save_total_limit=100,
|
164 |
load_best_model_at_end=True if VAL_SET_SIZE > 0 else False,
|
165 |
ddp_find_unused_parameters=False if ddp else None,
|
166 |
+
|
167 |
),
|
168 |
data_collator=transformers.DataCollatorForLanguageModeling(tokenizer, mlm=False),
|
169 |
)
|