Commit
·
40911be
1
Parent(s):
722813d
Upload 4 files
Browse files- UN-37/checkpoints/epoch=1-step=34331.ckpt +3 -0
- UN-37/hydra/config.yaml +184 -0
- UN-37/hydra/hydra.yaml +185 -0
- UN-37/hydra/overrides.yaml +17 -0
UN-37/checkpoints/epoch=1-step=34331.ckpt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5c6c2a473fcda127eaae3289905fb43b881bc93db0e422936be0b705320bc816
|
3 |
+
size 1020004119
|
UN-37/hydra/config.yaml
ADDED
@@ -0,0 +1,184 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
model:
|
2 |
+
optimizer:
|
3 |
+
_target_: transformers.AdamW
|
4 |
+
lr: 2.0e-05
|
5 |
+
betas:
|
6 |
+
- 0.9
|
7 |
+
- 0.98
|
8 |
+
eps: 1.0e-08
|
9 |
+
weight_decay: 0.0
|
10 |
+
correct_bias: false
|
11 |
+
scheduler:
|
12 |
+
lr_scheduler: linear_with_warmup
|
13 |
+
warmup_updates: 0.1
|
14 |
+
_target_: src.model.lm.LanguageModel
|
15 |
+
model: lm
|
16 |
+
arch: google/bigbird-roberta-base
|
17 |
+
dataset: ${data.dataset}
|
18 |
+
num_freeze_layers: 0
|
19 |
+
freeze_epochs: -1
|
20 |
+
expl_reg: true
|
21 |
+
train_topk:
|
22 |
+
- 1
|
23 |
+
- 5
|
24 |
+
- 10
|
25 |
+
- 20
|
26 |
+
- 50
|
27 |
+
eval_topk:
|
28 |
+
- 1
|
29 |
+
- 5
|
30 |
+
- 10
|
31 |
+
- 20
|
32 |
+
- 50
|
33 |
+
expl_reg_freq: 1
|
34 |
+
task_wt: 1.0
|
35 |
+
comp_criterion: margin
|
36 |
+
comp_margin: 1.0
|
37 |
+
comp_target: false
|
38 |
+
comp_wt: 0.5
|
39 |
+
suff_criterion: margin
|
40 |
+
suff_margin: 0.1
|
41 |
+
suff_target: false
|
42 |
+
suff_wt: 0.5
|
43 |
+
log_odds: false
|
44 |
+
log_odds_target: false
|
45 |
+
plaus_criterion: bce
|
46 |
+
plaus_margin: 0.1
|
47 |
+
plaus_wt: 0.5
|
48 |
+
explainer_type: lm
|
49 |
+
expl_head_type: linear
|
50 |
+
expl_head_mlp_hidden_dim: null
|
51 |
+
expl_head_mlp_hidden_layers: null
|
52 |
+
expl_head_mlp_dropout: 0.0
|
53 |
+
expl_head_mlp_layernorm: false
|
54 |
+
attr_algo: null
|
55 |
+
attr_pooling: null
|
56 |
+
attr_mlp_hidden_dim: null
|
57 |
+
attr_mlp_hidden_layers: null
|
58 |
+
attr_mlp_dropout: 0.0
|
59 |
+
attr_mlp_layernorm: false
|
60 |
+
ig_steps: 3
|
61 |
+
internal_batch_size: null
|
62 |
+
return_convergence_delta: false
|
63 |
+
gradshap_n_samples: null
|
64 |
+
gradshap_stdevs: null
|
65 |
+
fresh: false
|
66 |
+
fresh_extractor: null
|
67 |
+
l2e: false
|
68 |
+
l2e_wt: 0.0
|
69 |
+
l2e_criterion: ce
|
70 |
+
l2e_classes: 5
|
71 |
+
a2r: false
|
72 |
+
a2r_wt: 0.0
|
73 |
+
a2r_criterion: null
|
74 |
+
a2r_task_out: null
|
75 |
+
save_outputs: false
|
76 |
+
exp_id: null
|
77 |
+
measure_attrs_runtime: false
|
78 |
+
data:
|
79 |
+
_target_: src.data.data.DataModule
|
80 |
+
dataset: esnli
|
81 |
+
num_workers: ${setup.num_workers}
|
82 |
+
data_path: ${data_dir}/${.dataset}/${model.arch}/
|
83 |
+
train_batch_size: ${setup.train_batch_size}
|
84 |
+
eval_batch_size: ${setup.eval_batch_size}
|
85 |
+
eff_train_batch_size: ${setup.eff_train_batch_size}
|
86 |
+
mode: max
|
87 |
+
num_train: null
|
88 |
+
num_dev: null
|
89 |
+
num_test: null
|
90 |
+
num_train_seed: 0
|
91 |
+
num_dev_seed: 0
|
92 |
+
num_test_seed: 0
|
93 |
+
pct_train_rationales: null
|
94 |
+
pct_train_rationales_seed: 0
|
95 |
+
train_rationales_batch_factor: 2.0
|
96 |
+
neg_weight: 1
|
97 |
+
fresh_exp_id: null
|
98 |
+
fresh_attr_algo: null
|
99 |
+
fresh_topk: null
|
100 |
+
l2e_exp_id: null
|
101 |
+
l2e_attr_algo: null
|
102 |
+
logger:
|
103 |
+
_target_: src.utils.logging.get_neptune_logger
|
104 |
+
logger: neptune
|
105 |
+
log_db: manual_runs
|
106 |
+
tag_attrs:
|
107 |
+
- ${data.dataset}
|
108 |
+
- ${model.model}
|
109 |
+
- ${model.arch}
|
110 |
+
name: test
|
111 |
+
offline: false
|
112 |
+
project_name: shared.workspace/UNIREX
|
113 |
+
setup:
|
114 |
+
train_batch_size: 16
|
115 |
+
eval_batch_size: 1
|
116 |
+
accumulate_grad_batches: 2
|
117 |
+
eff_train_batch_size: 32
|
118 |
+
num_workers: 0
|
119 |
+
precision: 16
|
120 |
+
training:
|
121 |
+
evaluate_ckpt: false
|
122 |
+
finetune_ckpt: false
|
123 |
+
ckpt_path: null
|
124 |
+
eval_splits: all
|
125 |
+
train_shuffle: true
|
126 |
+
patience: 5
|
127 |
+
trainer:
|
128 |
+
_target_: pytorch_lightning.Trainer
|
129 |
+
checkpoint_callback: true
|
130 |
+
default_root_dir: null
|
131 |
+
gradient_clip_val: 0
|
132 |
+
process_position: 0
|
133 |
+
num_nodes: 1
|
134 |
+
num_processes: 1
|
135 |
+
gpus: -1
|
136 |
+
auto_select_gpus: true
|
137 |
+
tpu_cores: null
|
138 |
+
log_gpu_memory: null
|
139 |
+
progress_bar_refresh_rate: 1
|
140 |
+
overfit_batches: 0.0
|
141 |
+
track_grad_norm: -1
|
142 |
+
check_val_every_n_epoch: 1
|
143 |
+
fast_dev_run: false
|
144 |
+
accumulate_grad_batches: ${setup.accumulate_grad_batches}
|
145 |
+
max_epochs: 10
|
146 |
+
min_epochs: 1
|
147 |
+
max_steps: null
|
148 |
+
min_steps: null
|
149 |
+
limit_train_batches: 1.0
|
150 |
+
limit_val_batches: 1.0
|
151 |
+
limit_test_batches: 1.0
|
152 |
+
val_check_interval: 1.0
|
153 |
+
flush_logs_every_n_steps: 1000
|
154 |
+
log_every_n_steps: 1000
|
155 |
+
accelerator: null
|
156 |
+
sync_batchnorm: false
|
157 |
+
precision: ${setup.precision}
|
158 |
+
weights_summary: top
|
159 |
+
weights_save_path: null
|
160 |
+
num_sanity_val_steps: 0
|
161 |
+
truncated_bptt_steps: null
|
162 |
+
resume_from_checkpoint: null
|
163 |
+
profiler: null
|
164 |
+
benchmark: true
|
165 |
+
deterministic: true
|
166 |
+
reload_dataloaders_every_epoch: false
|
167 |
+
auto_lr_find: false
|
168 |
+
replace_sampler_ddp: true
|
169 |
+
terminate_on_nan: false
|
170 |
+
auto_scale_batch_size: false
|
171 |
+
prepare_data_per_node: true
|
172 |
+
plugins: null
|
173 |
+
amp_backend: native
|
174 |
+
amp_level: O2
|
175 |
+
move_metrics_to_cpu: false
|
176 |
+
seed: 0
|
177 |
+
debug: false
|
178 |
+
work_dir: ${hydra:runtime.cwd}
|
179 |
+
data_dir: ${work_dir}/../data
|
180 |
+
log_dir: ${work_dir}/../logs
|
181 |
+
save_dir: ${work_dir}/../save
|
182 |
+
save_checkpoint: true
|
183 |
+
save_rand_checkpoint: false
|
184 |
+
early_stopping: true
|
UN-37/hydra/hydra.yaml
ADDED
@@ -0,0 +1,185 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
hydra:
|
2 |
+
run:
|
3 |
+
dir: ${log_dir}/runs/${now:%Y-%m-%d}/${now:%H-%M-%S}
|
4 |
+
sweep:
|
5 |
+
dir: ../logs/multiruns/${now:%Y-%m-%d}/${now:%H-%M-%S}
|
6 |
+
subdir: ${hydra.job.num}
|
7 |
+
launcher:
|
8 |
+
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
|
9 |
+
sweeper:
|
10 |
+
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
|
11 |
+
max_batch_size: null
|
12 |
+
help:
|
13 |
+
app_name: ${hydra.job.name}
|
14 |
+
header: '${hydra.help.app_name} is powered by Hydra.
|
15 |
+
|
16 |
+
'
|
17 |
+
footer: 'Powered by Hydra (https://hydra.cc)
|
18 |
+
|
19 |
+
Use --hydra-help to view Hydra specific help
|
20 |
+
|
21 |
+
'
|
22 |
+
template: '${hydra.help.header}
|
23 |
+
|
24 |
+
== Configuration groups ==
|
25 |
+
|
26 |
+
Compose your configuration from those groups (group=option)
|
27 |
+
|
28 |
+
|
29 |
+
$APP_CONFIG_GROUPS
|
30 |
+
|
31 |
+
|
32 |
+
== Config ==
|
33 |
+
|
34 |
+
Override anything in the config (foo.bar=value)
|
35 |
+
|
36 |
+
|
37 |
+
$CONFIG
|
38 |
+
|
39 |
+
|
40 |
+
${hydra.help.footer}
|
41 |
+
|
42 |
+
'
|
43 |
+
hydra_help:
|
44 |
+
template: 'Hydra (${hydra.runtime.version})
|
45 |
+
|
46 |
+
See https://hydra.cc for more info.
|
47 |
+
|
48 |
+
|
49 |
+
== Flags ==
|
50 |
+
|
51 |
+
$FLAGS_HELP
|
52 |
+
|
53 |
+
|
54 |
+
== Configuration groups ==
|
55 |
+
|
56 |
+
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
|
57 |
+
to command line)
|
58 |
+
|
59 |
+
|
60 |
+
$HYDRA_CONFIG_GROUPS
|
61 |
+
|
62 |
+
|
63 |
+
Use ''--cfg hydra'' to Show the Hydra config.
|
64 |
+
|
65 |
+
'
|
66 |
+
hydra_help: ???
|
67 |
+
hydra_logging:
|
68 |
+
version: 1
|
69 |
+
formatters:
|
70 |
+
colorlog:
|
71 |
+
(): colorlog.ColoredFormatter
|
72 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
|
73 |
+
handlers:
|
74 |
+
console:
|
75 |
+
class: logging.StreamHandler
|
76 |
+
formatter: colorlog
|
77 |
+
stream: ext://sys.stdout
|
78 |
+
root:
|
79 |
+
level: INFO
|
80 |
+
handlers:
|
81 |
+
- console
|
82 |
+
disable_existing_loggers: false
|
83 |
+
job_logging:
|
84 |
+
version: 1
|
85 |
+
formatters:
|
86 |
+
simple:
|
87 |
+
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
|
88 |
+
colorlog:
|
89 |
+
(): colorlog.ColoredFormatter
|
90 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
|
91 |
+
- %(message)s'
|
92 |
+
log_colors:
|
93 |
+
DEBUG: purple
|
94 |
+
INFO: green
|
95 |
+
WARNING: yellow
|
96 |
+
ERROR: red
|
97 |
+
CRITICAL: red
|
98 |
+
handlers:
|
99 |
+
console:
|
100 |
+
class: logging.StreamHandler
|
101 |
+
formatter: colorlog
|
102 |
+
stream: ext://sys.stdout
|
103 |
+
file:
|
104 |
+
class: logging.FileHandler
|
105 |
+
formatter: simple
|
106 |
+
filename: ${hydra.job.name}.log
|
107 |
+
root:
|
108 |
+
level: INFO
|
109 |
+
handlers:
|
110 |
+
- console
|
111 |
+
- file
|
112 |
+
disable_existing_loggers: false
|
113 |
+
env: {}
|
114 |
+
searchpath: []
|
115 |
+
callbacks: {}
|
116 |
+
output_subdir: .hydra
|
117 |
+
overrides:
|
118 |
+
hydra: []
|
119 |
+
task:
|
120 |
+
- save_checkpoint=True
|
121 |
+
- data=esnli
|
122 |
+
- model=expl_reg
|
123 |
+
- model.explainer_type=lm
|
124 |
+
- model.expl_head_type=linear
|
125 |
+
- model.task_wt=1.0
|
126 |
+
- model.comp_wt=0.5
|
127 |
+
- model.suff_wt=0.5
|
128 |
+
- model.plaus_wt=0.5
|
129 |
+
- model.optimizer.lr=2e-05
|
130 |
+
- setup.train_batch_size=16
|
131 |
+
- setup.accumulate_grad_batches=2
|
132 |
+
- setup.eff_train_batch_size=32
|
133 |
+
- setup.eval_batch_size=1
|
134 |
+
- setup.num_workers=0
|
135 |
+
- seed=0
|
136 |
+
- model.save_outputs=False
|
137 |
+
job:
|
138 |
+
name: main
|
139 |
+
override_dirname: data=esnli,model.comp_wt=0.5,model.expl_head_type=linear,model.explainer_type=lm,model.optimizer.lr=2e-05,model.plaus_wt=0.5,model.save_outputs=False,model.suff_wt=0.5,model.task_wt=1.0,model=expl_reg,save_checkpoint=True,seed=0,setup.accumulate_grad_batches=2,setup.eff_train_batch_size=32,setup.eval_batch_size=1,setup.num_workers=0,setup.train_batch_size=16
|
140 |
+
id: '0'
|
141 |
+
num: 0
|
142 |
+
config_name: config
|
143 |
+
env_set: {}
|
144 |
+
env_copy: []
|
145 |
+
config:
|
146 |
+
override_dirname:
|
147 |
+
kv_sep: '='
|
148 |
+
item_sep: ','
|
149 |
+
exclude_keys: []
|
150 |
+
runtime:
|
151 |
+
version: 1.1.0
|
152 |
+
cwd: /home/pasquale/workspace/UNIREX2
|
153 |
+
config_sources:
|
154 |
+
- path: hydra.conf
|
155 |
+
schema: pkg
|
156 |
+
provider: hydra
|
157 |
+
- path: /home/pasquale/workspace/UNIREX2/configs
|
158 |
+
schema: file
|
159 |
+
provider: main
|
160 |
+
- path: hydra_plugins.hydra_colorlog.conf
|
161 |
+
schema: pkg
|
162 |
+
provider: hydra-colorlog
|
163 |
+
- path: ''
|
164 |
+
schema: structured
|
165 |
+
provider: schema
|
166 |
+
choices:
|
167 |
+
trainer: defaults
|
168 |
+
training: base
|
169 |
+
hydra: default
|
170 |
+
setup: a100
|
171 |
+
logger: neptune
|
172 |
+
data: esnli
|
173 |
+
model: expl_reg
|
174 |
+
model/scheduler: linear_with_warmup
|
175 |
+
model/optimizer: hf_adamw
|
176 |
+
hydra/env: default
|
177 |
+
hydra/callbacks: null
|
178 |
+
hydra/job_logging: colorlog
|
179 |
+
hydra/hydra_logging: colorlog
|
180 |
+
hydra/hydra_help: default
|
181 |
+
hydra/help: default
|
182 |
+
hydra/sweeper: basic
|
183 |
+
hydra/launcher: basic
|
184 |
+
hydra/output: default
|
185 |
+
verbose: false
|
UN-37/hydra/overrides.yaml
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
- save_checkpoint=True
|
2 |
+
- data=esnli
|
3 |
+
- model=expl_reg
|
4 |
+
- model.explainer_type=lm
|
5 |
+
- model.expl_head_type=linear
|
6 |
+
- model.task_wt=1.0
|
7 |
+
- model.comp_wt=0.5
|
8 |
+
- model.suff_wt=0.5
|
9 |
+
- model.plaus_wt=0.5
|
10 |
+
- model.optimizer.lr=2e-05
|
11 |
+
- setup.train_batch_size=16
|
12 |
+
- setup.accumulate_grad_batches=2
|
13 |
+
- setup.eff_train_batch_size=32
|
14 |
+
- setup.eval_batch_size=1
|
15 |
+
- setup.num_workers=0
|
16 |
+
- seed=0
|
17 |
+
- model.save_outputs=False
|