test / config.json
harrisonvanderbyl's picture
Update config.json
8b78505 verified
raw
history blame
246 Bytes
{
"_name_or_path": "harrisonvanderbyl/test",
"architectures": [
"LiquidHeadModelForCausalLM"
],
"vocab_size": 2,
"num_attention_heads":64,
"hidden_size": 4096,
"embedding_mode": true,
"num_hidden_layers": 1,
"head_dim": 64
}