tyleryzhu commited on
Commit
461333d
1 Parent(s): 934222c

Upload folder using huggingface_hub

Browse files
dinov2-single/config.json ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "align_stage_components": [
4
+ "download/videollava/valley_llavaimage.json",
5
+ "download/videollava"
6
+ ],
7
+ "dataset_id": "videollava",
8
+ "dataset_root_dir": "data",
9
+ "finetune_stage_components": [
10
+ "download/videollava/videochatgpt_llavaimage_tune.json",
11
+ "download/videollava"
12
+ ],
13
+ "type": "videollava"
14
+ },
15
+ "dataset_class": "FinetuneVideoDataset",
16
+ "hf_token": ".hf_token",
17
+ "model": {
18
+ "align_epochs": 1,
19
+ "align_global_batch_size": 256,
20
+ "align_learning_rate": 0.001,
21
+ "align_lr_scheduler_type": "linear-warmup+cosine-decay",
22
+ "align_max_grad_norm": 1.0,
23
+ "align_max_steps": null,
24
+ "align_per_device_batch_size": 16,
25
+ "align_train_strategy": "fsdp-shard-grad-op",
26
+ "align_warmup_ratio": 0.03,
27
+ "align_weight_decay": 0.0,
28
+ "arch_specifier": "no-align+linear",
29
+ "enable_gradient_checkpointing": true,
30
+ "enable_mixed_precision_training": true,
31
+ "feature_fusion": "first",
32
+ "finetune_epochs": 1,
33
+ "finetune_global_batch_size": 128,
34
+ "finetune_learning_rate": 2e-05,
35
+ "finetune_lr_scheduler_type": "linear-warmup+cosine-decay",
36
+ "finetune_max_grad_norm": 1.0,
37
+ "finetune_max_steps": null,
38
+ "finetune_per_device_batch_size": 4,
39
+ "finetune_train_strategy": "fsdp-full-shard",
40
+ "finetune_warmup_ratio": 0.03,
41
+ "finetune_weight_decay": 0.1,
42
+ "image_resize_strategy": "resize-naive",
43
+ "llm_backbone_id": "llama2-7b-pure",
44
+ "llm_max_length": 2048,
45
+ "model_id": "dinov2",
46
+ "num_frames": [
47
+ 16
48
+ ],
49
+ "projector_token_length": 64,
50
+ "reduce_in_full_precision": false,
51
+ "type": "dinov2-single",
52
+ "video_backbone_ids": [
53
+ "dinov2-video-all-token-with-cls"
54
+ ],
55
+ "visual_feature_length": 4096
56
+ },
57
+ "pretrained_checkpoint": null,
58
+ "run_id": "dinov2-single",
59
+ "run_root_dir": "runs",
60
+ "seed": 7,
61
+ "slurm_id": "21413390",
62
+ "stage": "finetune",
63
+ "trackers": [
64
+ "jsonl",
65
+ "wandb"
66
+ ]
67
+ }
dinov2-single/config.yaml ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ dataset:
2
+ align_stage_components:
3
+ - download/videollava/valley_llavaimage.json
4
+ - download/videollava
5
+ dataset_id: videollava
6
+ dataset_root_dir: data
7
+ finetune_stage_components:
8
+ - download/videollava/videochatgpt_llavaimage_tune.json
9
+ - download/videollava
10
+ type: videollava
11
+ dataset_class: FinetuneVideoDataset
12
+ hf_token: .hf_token
13
+ model:
14
+ align_epochs: 1
15
+ align_global_batch_size: 256
16
+ align_learning_rate: 0.001
17
+ align_lr_scheduler_type: linear-warmup+cosine-decay
18
+ align_max_grad_norm: 1.0
19
+ align_max_steps: null
20
+ align_per_device_batch_size: 16
21
+ align_train_strategy: fsdp-shard-grad-op
22
+ align_warmup_ratio: 0.03
23
+ align_weight_decay: 0.0
24
+ arch_specifier: no-align+linear
25
+ enable_gradient_checkpointing: true
26
+ enable_mixed_precision_training: true
27
+ feature_fusion: first
28
+ finetune_epochs: 1
29
+ finetune_global_batch_size: 128
30
+ finetune_learning_rate: 2.0e-05
31
+ finetune_lr_scheduler_type: linear-warmup+cosine-decay
32
+ finetune_max_grad_norm: 1.0
33
+ finetune_max_steps: null
34
+ finetune_per_device_batch_size: 4
35
+ finetune_train_strategy: fsdp-full-shard
36
+ finetune_warmup_ratio: 0.03
37
+ finetune_weight_decay: 0.1
38
+ image_resize_strategy: resize-naive
39
+ llm_backbone_id: llama2-7b-pure
40
+ llm_max_length: 2048
41
+ model_id: dinov2
42
+ num_frames:
43
+ - 16
44
+ projector_token_length: 64
45
+ reduce_in_full_precision: false
46
+ type: dinov2-single
47
+ video_backbone_ids:
48
+ - dinov2-video-all-token-with-cls
49
+ visual_feature_length: 4096
50
+ pretrained_checkpoint: null
51
+ run_id: dinov2-single
52
+ run_root_dir: runs
53
+ seed: 7
54
+ slurm_id: '21413390'
55
+ stage: finetune
56
+ trackers:
57
+ - jsonl
58
+ - wandb