Reza-Madani commited on
Commit
40911be
·
1 Parent(s): 722813d

Upload 4 files

Browse files
UN-37/checkpoints/epoch=1-step=34331.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c6c2a473fcda127eaae3289905fb43b881bc93db0e422936be0b705320bc816
3
+ size 1020004119
UN-37/hydra/config.yaml ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ optimizer:
3
+ _target_: transformers.AdamW
4
+ lr: 2.0e-05
5
+ betas:
6
+ - 0.9
7
+ - 0.98
8
+ eps: 1.0e-08
9
+ weight_decay: 0.0
10
+ correct_bias: false
11
+ scheduler:
12
+ lr_scheduler: linear_with_warmup
13
+ warmup_updates: 0.1
14
+ _target_: src.model.lm.LanguageModel
15
+ model: lm
16
+ arch: google/bigbird-roberta-base
17
+ dataset: ${data.dataset}
18
+ num_freeze_layers: 0
19
+ freeze_epochs: -1
20
+ expl_reg: true
21
+ train_topk:
22
+ - 1
23
+ - 5
24
+ - 10
25
+ - 20
26
+ - 50
27
+ eval_topk:
28
+ - 1
29
+ - 5
30
+ - 10
31
+ - 20
32
+ - 50
33
+ expl_reg_freq: 1
34
+ task_wt: 1.0
35
+ comp_criterion: margin
36
+ comp_margin: 1.0
37
+ comp_target: false
38
+ comp_wt: 0.5
39
+ suff_criterion: margin
40
+ suff_margin: 0.1
41
+ suff_target: false
42
+ suff_wt: 0.5
43
+ log_odds: false
44
+ log_odds_target: false
45
+ plaus_criterion: bce
46
+ plaus_margin: 0.1
47
+ plaus_wt: 0.5
48
+ explainer_type: lm
49
+ expl_head_type: linear
50
+ expl_head_mlp_hidden_dim: null
51
+ expl_head_mlp_hidden_layers: null
52
+ expl_head_mlp_dropout: 0.0
53
+ expl_head_mlp_layernorm: false
54
+ attr_algo: null
55
+ attr_pooling: null
56
+ attr_mlp_hidden_dim: null
57
+ attr_mlp_hidden_layers: null
58
+ attr_mlp_dropout: 0.0
59
+ attr_mlp_layernorm: false
60
+ ig_steps: 3
61
+ internal_batch_size: null
62
+ return_convergence_delta: false
63
+ gradshap_n_samples: null
64
+ gradshap_stdevs: null
65
+ fresh: false
66
+ fresh_extractor: null
67
+ l2e: false
68
+ l2e_wt: 0.0
69
+ l2e_criterion: ce
70
+ l2e_classes: 5
71
+ a2r: false
72
+ a2r_wt: 0.0
73
+ a2r_criterion: null
74
+ a2r_task_out: null
75
+ save_outputs: false
76
+ exp_id: null
77
+ measure_attrs_runtime: false
78
+ data:
79
+ _target_: src.data.data.DataModule
80
+ dataset: esnli
81
+ num_workers: ${setup.num_workers}
82
+ data_path: ${data_dir}/${.dataset}/${model.arch}/
83
+ train_batch_size: ${setup.train_batch_size}
84
+ eval_batch_size: ${setup.eval_batch_size}
85
+ eff_train_batch_size: ${setup.eff_train_batch_size}
86
+ mode: max
87
+ num_train: null
88
+ num_dev: null
89
+ num_test: null
90
+ num_train_seed: 0
91
+ num_dev_seed: 0
92
+ num_test_seed: 0
93
+ pct_train_rationales: null
94
+ pct_train_rationales_seed: 0
95
+ train_rationales_batch_factor: 2.0
96
+ neg_weight: 1
97
+ fresh_exp_id: null
98
+ fresh_attr_algo: null
99
+ fresh_topk: null
100
+ l2e_exp_id: null
101
+ l2e_attr_algo: null
102
+ logger:
103
+ _target_: src.utils.logging.get_neptune_logger
104
+ logger: neptune
105
+ log_db: manual_runs
106
+ tag_attrs:
107
+ - ${data.dataset}
108
+ - ${model.model}
109
+ - ${model.arch}
110
+ name: test
111
+ offline: false
112
+ project_name: shared.workspace/UNIREX
113
+ setup:
114
+ train_batch_size: 16
115
+ eval_batch_size: 1
116
+ accumulate_grad_batches: 2
117
+ eff_train_batch_size: 32
118
+ num_workers: 0
119
+ precision: 16
120
+ training:
121
+ evaluate_ckpt: false
122
+ finetune_ckpt: false
123
+ ckpt_path: null
124
+ eval_splits: all
125
+ train_shuffle: true
126
+ patience: 5
127
+ trainer:
128
+ _target_: pytorch_lightning.Trainer
129
+ checkpoint_callback: true
130
+ default_root_dir: null
131
+ gradient_clip_val: 0
132
+ process_position: 0
133
+ num_nodes: 1
134
+ num_processes: 1
135
+ gpus: -1
136
+ auto_select_gpus: true
137
+ tpu_cores: null
138
+ log_gpu_memory: null
139
+ progress_bar_refresh_rate: 1
140
+ overfit_batches: 0.0
141
+ track_grad_norm: -1
142
+ check_val_every_n_epoch: 1
143
+ fast_dev_run: false
144
+ accumulate_grad_batches: ${setup.accumulate_grad_batches}
145
+ max_epochs: 10
146
+ min_epochs: 1
147
+ max_steps: null
148
+ min_steps: null
149
+ limit_train_batches: 1.0
150
+ limit_val_batches: 1.0
151
+ limit_test_batches: 1.0
152
+ val_check_interval: 1.0
153
+ flush_logs_every_n_steps: 1000
154
+ log_every_n_steps: 1000
155
+ accelerator: null
156
+ sync_batchnorm: false
157
+ precision: ${setup.precision}
158
+ weights_summary: top
159
+ weights_save_path: null
160
+ num_sanity_val_steps: 0
161
+ truncated_bptt_steps: null
162
+ resume_from_checkpoint: null
163
+ profiler: null
164
+ benchmark: true
165
+ deterministic: true
166
+ reload_dataloaders_every_epoch: false
167
+ auto_lr_find: false
168
+ replace_sampler_ddp: true
169
+ terminate_on_nan: false
170
+ auto_scale_batch_size: false
171
+ prepare_data_per_node: true
172
+ plugins: null
173
+ amp_backend: native
174
+ amp_level: O2
175
+ move_metrics_to_cpu: false
176
+ seed: 0
177
+ debug: false
178
+ work_dir: ${hydra:runtime.cwd}
179
+ data_dir: ${work_dir}/../data
180
+ log_dir: ${work_dir}/../logs
181
+ save_dir: ${work_dir}/../save
182
+ save_checkpoint: true
183
+ save_rand_checkpoint: false
184
+ early_stopping: true
UN-37/hydra/hydra.yaml ADDED
@@ -0,0 +1,185 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: ${log_dir}/runs/${now:%Y-%m-%d}/${now:%H-%M-%S}
4
+ sweep:
5
+ dir: ../logs/multiruns/${now:%Y-%m-%d}/${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ help:
13
+ app_name: ${hydra.job.name}
14
+ header: '${hydra.help.app_name} is powered by Hydra.
15
+
16
+ '
17
+ footer: 'Powered by Hydra (https://hydra.cc)
18
+
19
+ Use --hydra-help to view Hydra specific help
20
+
21
+ '
22
+ template: '${hydra.help.header}
23
+
24
+ == Configuration groups ==
25
+
26
+ Compose your configuration from those groups (group=option)
27
+
28
+
29
+ $APP_CONFIG_GROUPS
30
+
31
+
32
+ == Config ==
33
+
34
+ Override anything in the config (foo.bar=value)
35
+
36
+
37
+ $CONFIG
38
+
39
+
40
+ ${hydra.help.footer}
41
+
42
+ '
43
+ hydra_help:
44
+ template: 'Hydra (${hydra.runtime.version})
45
+
46
+ See https://hydra.cc for more info.
47
+
48
+
49
+ == Flags ==
50
+
51
+ $FLAGS_HELP
52
+
53
+
54
+ == Configuration groups ==
55
+
56
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
57
+ to command line)
58
+
59
+
60
+ $HYDRA_CONFIG_GROUPS
61
+
62
+
63
+ Use ''--cfg hydra'' to Show the Hydra config.
64
+
65
+ '
66
+ hydra_help: ???
67
+ hydra_logging:
68
+ version: 1
69
+ formatters:
70
+ colorlog:
71
+ (): colorlog.ColoredFormatter
72
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: colorlog
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ disable_existing_loggers: false
83
+ job_logging:
84
+ version: 1
85
+ formatters:
86
+ simple:
87
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
88
+ colorlog:
89
+ (): colorlog.ColoredFormatter
90
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
91
+ - %(message)s'
92
+ log_colors:
93
+ DEBUG: purple
94
+ INFO: green
95
+ WARNING: yellow
96
+ ERROR: red
97
+ CRITICAL: red
98
+ handlers:
99
+ console:
100
+ class: logging.StreamHandler
101
+ formatter: colorlog
102
+ stream: ext://sys.stdout
103
+ file:
104
+ class: logging.FileHandler
105
+ formatter: simple
106
+ filename: ${hydra.job.name}.log
107
+ root:
108
+ level: INFO
109
+ handlers:
110
+ - console
111
+ - file
112
+ disable_existing_loggers: false
113
+ env: {}
114
+ searchpath: []
115
+ callbacks: {}
116
+ output_subdir: .hydra
117
+ overrides:
118
+ hydra: []
119
+ task:
120
+ - save_checkpoint=True
121
+ - data=esnli
122
+ - model=expl_reg
123
+ - model.explainer_type=lm
124
+ - model.expl_head_type=linear
125
+ - model.task_wt=1.0
126
+ - model.comp_wt=0.5
127
+ - model.suff_wt=0.5
128
+ - model.plaus_wt=0.5
129
+ - model.optimizer.lr=2e-05
130
+ - setup.train_batch_size=16
131
+ - setup.accumulate_grad_batches=2
132
+ - setup.eff_train_batch_size=32
133
+ - setup.eval_batch_size=1
134
+ - setup.num_workers=0
135
+ - seed=0
136
+ - model.save_outputs=False
137
+ job:
138
+ name: main
139
+ override_dirname: data=esnli,model.comp_wt=0.5,model.expl_head_type=linear,model.explainer_type=lm,model.optimizer.lr=2e-05,model.plaus_wt=0.5,model.save_outputs=False,model.suff_wt=0.5,model.task_wt=1.0,model=expl_reg,save_checkpoint=True,seed=0,setup.accumulate_grad_batches=2,setup.eff_train_batch_size=32,setup.eval_batch_size=1,setup.num_workers=0,setup.train_batch_size=16
140
+ id: '0'
141
+ num: 0
142
+ config_name: config
143
+ env_set: {}
144
+ env_copy: []
145
+ config:
146
+ override_dirname:
147
+ kv_sep: '='
148
+ item_sep: ','
149
+ exclude_keys: []
150
+ runtime:
151
+ version: 1.1.0
152
+ cwd: /home/pasquale/workspace/UNIREX2
153
+ config_sources:
154
+ - path: hydra.conf
155
+ schema: pkg
156
+ provider: hydra
157
+ - path: /home/pasquale/workspace/UNIREX2/configs
158
+ schema: file
159
+ provider: main
160
+ - path: hydra_plugins.hydra_colorlog.conf
161
+ schema: pkg
162
+ provider: hydra-colorlog
163
+ - path: ''
164
+ schema: structured
165
+ provider: schema
166
+ choices:
167
+ trainer: defaults
168
+ training: base
169
+ hydra: default
170
+ setup: a100
171
+ logger: neptune
172
+ data: esnli
173
+ model: expl_reg
174
+ model/scheduler: linear_with_warmup
175
+ model/optimizer: hf_adamw
176
+ hydra/env: default
177
+ hydra/callbacks: null
178
+ hydra/job_logging: colorlog
179
+ hydra/hydra_logging: colorlog
180
+ hydra/hydra_help: default
181
+ hydra/help: default
182
+ hydra/sweeper: basic
183
+ hydra/launcher: basic
184
+ hydra/output: default
185
+ verbose: false
UN-37/hydra/overrides.yaml ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - save_checkpoint=True
2
+ - data=esnli
3
+ - model=expl_reg
4
+ - model.explainer_type=lm
5
+ - model.expl_head_type=linear
6
+ - model.task_wt=1.0
7
+ - model.comp_wt=0.5
8
+ - model.suff_wt=0.5
9
+ - model.plaus_wt=0.5
10
+ - model.optimizer.lr=2e-05
11
+ - setup.train_batch_size=16
12
+ - setup.accumulate_grad_batches=2
13
+ - setup.eff_train_batch_size=32
14
+ - setup.eval_batch_size=1
15
+ - setup.num_workers=0
16
+ - seed=0
17
+ - model.save_outputs=False