gustavoaq commited on
Commit
e907308
·
1 Parent(s): ea53189

Update finetune.py

Browse files
Files changed (1) hide show
  1. finetune.py +13 -2
finetune.py CHANGED
@@ -15,9 +15,19 @@ from peft import (
15
  get_peft_model_state_dict,
16
  )
17
 
18
-
19
- # Parameters
20
  HF_TOKEN = os.environ.get("TRL_TOKEN", None)
 
 
 
 
 
 
 
 
 
 
 
 
21
  MICRO_BATCH_SIZE = 16
22
  BATCH_SIZE = 16
23
  size = "7b"
@@ -153,6 +163,7 @@ trainer = transformers.Trainer(
153
  save_total_limit=100,
154
  load_best_model_at_end=True if VAL_SET_SIZE > 0 else False,
155
  ddp_find_unused_parameters=False if ddp else None,
 
156
  ),
157
  data_collator=transformers.DataCollatorForLanguageModeling(tokenizer, mlm=False),
158
  )
 
15
  get_peft_model_state_dict,
16
  )
17
 
 
 
18
  HF_TOKEN = os.environ.get("TRL_TOKEN", None)
19
+ if HF_TOKEN:
20
+ try:
21
+ shutil.rmtree("./data/")
22
+ except:
23
+ pass
24
+
25
+ repo = Repository(
26
+ local_dir="./data/", clone_from="gustavoaq/llama_ft", use_auth_token=HF_TOKEN, repo_type="models"
27
+ )
28
+ repo.git_pull()
29
+ # Parameters
30
+
31
  MICRO_BATCH_SIZE = 16
32
  BATCH_SIZE = 16
33
  size = "7b"
 
163
  save_total_limit=100,
164
  load_best_model_at_end=True if VAL_SET_SIZE > 0 else False,
165
  ddp_find_unused_parameters=False if ddp else None,
166
+
167
  ),
168
  data_collator=transformers.DataCollatorForLanguageModeling(tokenizer, mlm=False),
169
  )