ZhangYuanhan commited on
Commit
a156b3d
1 Parent(s): c9f2a86

Upload config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. config.json +5 -21
config.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
- "_name_or_path": "",
3
  "add_faster_video": false,
4
  "add_time_instruction": true,
5
- "vocab_size": 152064,
6
  "architectures": [
7
  "LlavaQwenForCausalLM"
8
  ],
9
  "attention_dropout": 0.0,
 
10
  "bos_token_id": 151643,
 
11
  "eos_token_id": 151645,
12
- "faster_token_stride": 10,
13
- "force_sample": true,
14
  "hidden_act": "silu",
15
  "hidden_size": 8192,
16
  "ignore_index": -100,
@@ -193,11 +193,8 @@
193
  "rope_scaling": null,
194
  "rope_theta": 1000000.0,
195
  "sliding_window": 131072,
196
- "text_config": {
197
- "model_type": "llama"
198
- },
199
  "tie_word_embeddings": false,
200
- "tokenizer_model_max_length": 32768,
201
  "tokenizer_padding_side": "right",
202
  "torch_dtype": "bfloat16",
203
  "transformers_version": "4.40.0.dev0",
@@ -205,18 +202,5 @@
205
  "use_mm_proj": true,
206
  "use_pos_skipping": false,
207
  "use_sliding_window": false,
208
- "vision_config": {
209
- "hidden_size": 1024,
210
- "image_size": 336,
211
- "intermediate_size": 4096,
212
- "model_type": "clip_vision_model",
213
- "num_attention_heads": 16,
214
- "num_hidden_layers": 24,
215
- "patch_size": 14,
216
- "projection_dim": 768,
217
- "vocab_size": 32000
218
- },
219
- "vision_feature_layer": -2,
220
- "vision_feature_select_strategy": "default",
221
  "vision_tower_pretrained": null
222
  }
 
1
  {
2
+ "_name_or_path": "/mnt/bn/vl-research/checkpoints/onevision/llavanext-google_siglip-so400m-patch14-384-Qwen_Qwen2-72B-Instruct-final_to_ov_am9_july31v4/checkpoint-9000",
3
  "add_faster_video": false,
4
  "add_time_instruction": true,
5
+ "force_sample": true,
6
  "architectures": [
7
  "LlavaQwenForCausalLM"
8
  ],
9
  "attention_dropout": 0.0,
10
+ "vocab_size": 152064,
11
  "bos_token_id": 151643,
12
+ "delay_load": false,
13
  "eos_token_id": 151645,
 
 
14
  "hidden_act": "silu",
15
  "hidden_size": 8192,
16
  "ignore_index": -100,
 
193
  "rope_scaling": null,
194
  "rope_theta": 1000000.0,
195
  "sliding_window": 131072,
 
 
 
196
  "tie_word_embeddings": false,
197
+ "tokenizer_model_max_length": 12768,
198
  "tokenizer_padding_side": "right",
199
  "torch_dtype": "bfloat16",
200
  "transformers_version": "4.40.0.dev0",
 
202
  "use_mm_proj": true,
203
  "use_pos_skipping": false,
204
  "use_sliding_window": false,
 
 
 
 
 
 
 
 
 
 
 
 
 
205
  "vision_tower_pretrained": null
206
  }