Spaces:
Paused
Paused
Update finetune.py
Browse files- finetune.py +4 -1
finetune.py
CHANGED
@@ -17,7 +17,7 @@ from peft import (
|
|
17 |
|
18 |
|
19 |
# Parameters
|
20 |
-
MICRO_BATCH_SIZE =
|
21 |
BATCH_SIZE = 16
|
22 |
size = "7b"
|
23 |
GRADIENT_ACCUMULATION_STEPS = BATCH_SIZE // MICRO_BATCH_SIZE
|
@@ -152,6 +152,8 @@ trainer = transformers.Trainer(
|
|
152 |
save_total_limit=100,
|
153 |
load_best_model_at_end=True if VAL_SET_SIZE > 0 else False,
|
154 |
ddp_find_unused_parameters=False if ddp else None,
|
|
|
|
|
155 |
),
|
156 |
data_collator=transformers.DataCollatorForLanguageModeling(tokenizer, mlm=False),
|
157 |
)
|
@@ -163,5 +165,6 @@ model.state_dict = (
|
|
163 |
).__get__(model, type(model))
|
164 |
|
165 |
trainer.train()
|
|
|
166 |
|
167 |
model.save_pretrained(OUTPUT_DIR)
|
|
|
17 |
|
18 |
|
19 |
# Parameters
|
20 |
+
MICRO_BATCH_SIZE = 16
|
21 |
BATCH_SIZE = 16
|
22 |
size = "7b"
|
23 |
GRADIENT_ACCUMULATION_STEPS = BATCH_SIZE // MICRO_BATCH_SIZE
|
|
|
152 |
save_total_limit=100,
|
153 |
load_best_model_at_end=True if VAL_SET_SIZE > 0 else False,
|
154 |
ddp_find_unused_parameters=False if ddp else None,
|
155 |
+
push_to_hub=True,
|
156 |
+
push_to_hub_model_id="gustavoaq/llama_ft"
|
157 |
),
|
158 |
data_collator=transformers.DataCollatorForLanguageModeling(tokenizer, mlm=False),
|
159 |
)
|
|
|
165 |
).__get__(model, type(model))
|
166 |
|
167 |
trainer.train()
|
168 |
+
trainer.push_to_hub()
|
169 |
|
170 |
model.save_pretrained(OUTPUT_DIR)
|