fukugawa's picture
Upload FlaxTransformerLMForCausalLM
7e9336f verified
raw
history blame contribute delete
733 Bytes
{
"architectures": [
"TransformerLMForCausalLM"
],
"attention_dropout_rate": 0.1,
"auto_map": {
"AutoConfig": "configuration_transformerlm.TransformerLMConfig",
"FlaxAutoModelForCausalLM": "modeling_transformerlm_flax.FlaxTransformerLMForCausalLM"
},
"bos_token_id": 50256,
"decode": true,
"deterministic": true,
"dropout_rate": 0.1,
"emb_dim": 768,
"eos_token_id": 50256,
"logits_via_embedding": false,
"max_len": 512,
"mlp_dim": 3072,
"model_type": "transformerlm",
"num_heads": 12,
"num_layers": 12,
"output_vocab_size": 30000,
"qkv_dim": 768,
"share_embeddings": false,
"tokenizer_class": "TransformerLMTokenizer",
"transformers_version": "4.39.0",
"vocab_size": 30000
}