File size: 1,081 Bytes
a56282a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
{
  "add_qkv_bias": true,
  "asr_adapter": "llamamlp",
  "attn_dropout": 0,
  "bias": false,
  "block_size": 2048,
  "force_align": false,
  "gelu_approximate": "none",
  "head_size": 64,
  "hf_config": {
    "name": "Qwen2-0.5B",
    "org": "Qwen"
  },
  "intermediate_size": 4864,
  "lm_head_bias": false,
  "mlp_class_name": "LLaMAMLP",
  "n_embd": 896,
  "n_expert": 0,
  "n_expert_per_token": 0,
  "n_head": 14,
  "n_layer": 24,
  "n_query_groups": 2,
  "name": "Qwen2-0.5B",
  "norm_class_name": "RMSNorm",
  "norm_eps": 0.000001,
  "padded_vocab_size": 181120,
  "padding_multiple": 512,
  "parallel_residual": false,
  "pos_type": "rope",
  "post_adapter": false,
  "post_adapter_layers": 6,
  "prompt_vocab_size": null,
  "rope_base": 1000000,
  "rope_condense_ratio": 1,
  "rotary_percentage": 1,
  "scale_embeddings": false,
  "shared_attention_norm": false,
  "tie_word_embeddings": true,
  "use_pretrain_phoneme_emb": false,
  "vocab_size": 50254,
  "text_vocab_size": 152000,
  "cat_audio_vocab_size": 29120,
  "audio_vocab_size": 4160,
  "whisper_adapter_dim": 768
}