{ "architectures": [ "Gemma2ForCausalLM" ], "module": "keras_nlp.src.models.gemma.gemma_backbone", "class_name": "GemmaBackbone", "model_type": "gemma", "name": "gemma_backbone", "trainable": true, "vocabulary_size": 256000, "num_layers": 18, "num_query_heads": 8, "num_key_value_heads": 1, "hidden_dim": 2048, "intermediate_dim": 32768, "head_dim": 256, "layer_norm_epsilon": 1e-06, "dropout": 0, "query_head_dim_normalize": true, "use_post_ffw_norm": false, "use_post_attention_norm": false, "final_logit_soft_cap": null, "attention_logit_soft_cap": null, "sliding_window_size": 4096, "use_sliding_window_attention": false, "registered_name": "keras_nlp>GemmaBackbone" }