{ "_name_or_path": "Alibaba-NLP/gte-multilingual-base", "architectures": [ "NewForSequenceClassification" ], "attention_probs_dropout_prob": 0.0, "auto_map": { "AutoConfig": "Alibaba-NLP/new-impl--configuration.NewConfig", "AutoModel": "Alibaba-NLP/new-impl--modeling.NewModel", "AutoModelForMaskedLM": "Alibaba-NLP/new-impl--modeling.NewForMaskedLM", "AutoModelForMultipleChoice": "Alibaba-NLP/new-impl--modeling.NewForMultipleChoice", "AutoModelForQuestionAnswering": "Alibaba-NLP/new-impl--modeling.NewForQuestionAnswering", "AutoModelForSequenceClassification": "Alibaba-NLP/new-impl--modeling.NewForSequenceClassification", "AutoModelForTokenClassification": "Alibaba-NLP/new-impl--modeling.NewForTokenClassification" }, "classifier_dropout": 0.0, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "H\u00f4 h\u1ea5p - Ph\u1ed5i", "1": "Chuy\u00ean khoa M\u1eaft", "2": "B\u1ec7nh Vi\u00eam gan", "3": "N\u1ed9i khoa", "4": "Th\u1ea7n kinh", "5": "Ti\u00eau ho\u00e1", "6": "Tim m\u1ea1ch", "7": "Th\u1eadn - Ti\u1ebft ni\u1ec7u", "8": "Ti\u1ec3u \u0111\u01b0\u1eddng - N\u1ed9i ti\u1ebft", "9": "C\u01a1 X\u01b0\u01a1ng Kh\u1edbp", "10": "Tai M\u0169i H\u1ecdng", "11": "Ch\u1ea5n th\u01b0\u01a1ng ch\u1ec9nh h\u00ecnh", "12": "Ung b\u01b0\u1edbu", "13": "Da li\u1ec5u", "14": "Nha khoa", "15": "Nam h\u1ecdc", "16": "Ph\u1ee5c h\u1ed3i ch\u1ee9c n\u0103ng", "17": "C\u1ed9t s\u1ed1ng", "18": "Nhi khoa", "19": "T\u1ea1o h\u00ecnh H\u00e0m M\u1eb7t", "20": "Ngo\u1ea1i th\u1ea7n kinh", "21": "Ph\u1eabu thu\u1eadt H\u1eadu m\u00f4n - Tr\u1ef1c tr\u00e0ng" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "B\u1ec7nh Vi\u00eam gan": 2, "Chuy\u00ean khoa M\u1eaft": 1, "Ch\u1ea5n th\u01b0\u01a1ng ch\u1ec9nh h\u00ecnh": 11, "C\u01a1 X\u01b0\u01a1ng Kh\u1edbp": 9, "C\u1ed9t s\u1ed1ng": 17, "Da li\u1ec5u": 13, "H\u00f4 h\u1ea5p - Ph\u1ed5i": 0, "Nam h\u1ecdc": 15, "Ngo\u1ea1i th\u1ea7n kinh": 20, "Nha khoa": 14, "Nhi khoa": 18, "N\u1ed9i khoa": 3, "Ph\u1eabu thu\u1eadt H\u1eadu m\u00f4n - Tr\u1ef1c tr\u00e0ng": 21, "Ph\u1ee5c h\u1ed3i ch\u1ee9c n\u0103ng": 16, "Tai M\u0169i H\u1ecdng": 10, "Th\u1ea7n kinh": 4, "Th\u1eadn - Ti\u1ebft ni\u1ec7u": 7, "Tim m\u1ea1ch": 6, "Ti\u00eau ho\u00e1": 5, "Ti\u1ec3u \u0111\u01b0\u1eddng - N\u1ed9i ti\u1ebft": 8, "T\u1ea1o h\u00ecnh H\u00e0m M\u1eb7t": 19, "Ung b\u01b0\u1edbu": 12 }, "layer_norm_eps": 1e-12, "layer_norm_type": "layer_norm", "logn_attention_clip1": false, "logn_attention_scale": false, "max_position_embeddings": 8192, "model_type": "new", "num_attention_heads": 12, "num_hidden_layers": 12, "pack_qkv": true, "pad_token_id": 1, "position_embedding_type": "rope", "problem_type": "single_label_classification", "rope_scaling": { "factor": 8.0, "type": "ntk" }, "rope_theta": 20000, "torch_dtype": "float32", "transformers_version": "4.47.1", "type_vocab_size": 1, "unpad_inputs": false, "use_memory_efficient_attention": false, "vocab_size": 250048 }