AjayP13 commited on
Commit
f0b1ec0
·
verified ·
1 Parent(s): 3ca22f9

Pushed by DataDreamer

Browse files

Update training_args.json

Files changed (1) hide show
  1. training_args.json +4 -4
training_args.json CHANGED
@@ -10,16 +10,16 @@
10
  "per_device_eval_batch_size": 16,
11
  "per_gpu_train_batch_size": null,
12
  "per_gpu_eval_batch_size": null,
13
- "gradient_accumulation_steps": 8,
14
  "eval_accumulation_steps": 1,
15
  "eval_delay": 0,
16
- "learning_rate": 0.001,
17
  "weight_decay": 0.01,
18
  "adam_beta1": 0.9,
19
  "adam_beta2": 0.999,
20
  "adam_epsilon": 1e-08,
21
  "max_grad_norm": 1.0,
22
- "num_train_epochs": 20,
23
  "max_steps": -1,
24
  "lr_scheduler_type": "linear",
25
  "lr_scheduler_kwargs": {},
@@ -28,7 +28,7 @@
28
  "log_level": "passive",
29
  "log_level_replica": "warning",
30
  "log_on_each_node": true,
31
- "logging_dir": "./output/train-wegmann--styledistance-model/_checkpoints/runs/Jul21_05-54-20_nlpgpu04.seas.upenn.edu",
32
  "logging_strategy": "steps",
33
  "logging_first_step": false,
34
  "logging_steps": 1,
 
10
  "per_device_eval_batch_size": 16,
11
  "per_gpu_train_batch_size": null,
12
  "per_gpu_eval_batch_size": null,
13
+ "gradient_accumulation_steps": 4,
14
  "eval_accumulation_steps": 1,
15
  "eval_delay": 0,
16
+ "learning_rate": 0.0001,
17
  "weight_decay": 0.01,
18
  "adam_beta1": 0.9,
19
  "adam_beta2": 0.999,
20
  "adam_epsilon": 1e-08,
21
  "max_grad_norm": 1.0,
22
+ "num_train_epochs": 10,
23
  "max_steps": -1,
24
  "lr_scheduler_type": "linear",
25
  "lr_scheduler_kwargs": {},
 
28
  "log_level": "passive",
29
  "log_level_replica": "warning",
30
  "log_on_each_node": true,
31
+ "logging_dir": "./output/train-wegmann--styledistance-model/_checkpoints/runs/Jul23_14-22-58_nlpgpu04.seas.upenn.edu",
32
  "logging_strategy": "steps",
33
  "logging_first_step": false,
34
  "logging_steps": 1,