--- language: - pt license: mit tags: - text-generation-inference - transformers - unsloth - phi base_model: unsloth/Phi-3-mini-4k-instruct-bnb-4bit datasets: - dominguesm/Canarim-Instruct-PTBR-Dataset model-index: - name: Phituguese_FP16 results: - task: type: text-generation name: Text Generation dataset: name: ENEM Challenge (No Images) type: eduagarcia/enem_challenge split: train args: num_few_shot: 3 metrics: - type: acc value: 49.97 name: accuracy source: url: https://huggingface.co./spaces/eduagarcia/open_pt_llm_leaderboard?query=GuiCas/Phituguese_FP16 name: Open Portuguese LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: BLUEX (No Images) type: eduagarcia-temp/BLUEX_without_images split: train args: num_few_shot: 3 metrics: - type: acc value: 43.25 name: accuracy source: url: https://huggingface.co./spaces/eduagarcia/open_pt_llm_leaderboard?query=GuiCas/Phituguese_FP16 name: Open Portuguese LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: OAB Exams type: eduagarcia/oab_exams split: train args: num_few_shot: 3 metrics: - type: acc value: 38.13 name: accuracy source: url: https://huggingface.co./spaces/eduagarcia/open_pt_llm_leaderboard?query=GuiCas/Phituguese_FP16 name: Open Portuguese LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: Assin2 RTE type: assin2 split: test args: num_few_shot: 15 metrics: - type: f1_macro value: 74.75 name: f1-macro source: url: https://huggingface.co./spaces/eduagarcia/open_pt_llm_leaderboard?query=GuiCas/Phituguese_FP16 name: Open Portuguese LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: Assin2 STS type: eduagarcia/portuguese_benchmark split: test args: num_few_shot: 15 metrics: - type: pearson value: 71.93 name: pearson source: url: https://huggingface.co./spaces/eduagarcia/open_pt_llm_leaderboard?query=GuiCas/Phituguese_FP16 name: Open Portuguese LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: FaQuAD NLI type: ruanchaves/faquad-nli split: test args: num_few_shot: 15 metrics: - type: f1_macro value: 43.97 name: f1-macro source: url: https://huggingface.co./spaces/eduagarcia/open_pt_llm_leaderboard?query=GuiCas/Phituguese_FP16 name: Open Portuguese LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: HateBR Binary type: ruanchaves/hatebr split: test args: num_few_shot: 25 metrics: - type: f1_macro value: 57.34 name: f1-macro source: url: https://huggingface.co./spaces/eduagarcia/open_pt_llm_leaderboard?query=GuiCas/Phituguese_FP16 name: Open Portuguese LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: PT Hate Speech Binary type: hate_speech_portuguese split: test args: num_few_shot: 25 metrics: - type: f1_macro value: 60.48 name: f1-macro source: url: https://huggingface.co./spaces/eduagarcia/open_pt_llm_leaderboard?query=GuiCas/Phituguese_FP16 name: Open Portuguese LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: tweetSentBR type: eduagarcia/tweetsentbr_fewshot split: test args: num_few_shot: 25 metrics: - type: f1_macro value: 61.11 name: f1-macro source: url: https://huggingface.co./spaces/eduagarcia/open_pt_llm_leaderboard?query=GuiCas/Phituguese_FP16 name: Open Portuguese LLM Leaderboard --- # Uploaded model - **Developed by:** GuiCas - **License:** mit - **Finetuned from model :** unsloth/Phi-3-mini-4k-instruct-bnb-4bit [](https://github.com/unslothai/unsloth) # Open Portuguese LLM Leaderboard Evaluation Results Detailed results can be found [here](https://huggingface.co./datasets/eduagarcia-temp/llm_pt_leaderboard_raw_results/tree/main/GuiCas/Phituguese_FP16) and on the [🚀 Open Portuguese LLM Leaderboard](https://huggingface.co./spaces/eduagarcia/open_pt_llm_leaderboard) | Metric | Value | |--------------------------|---------| |Average |**55.66**| |ENEM Challenge (No Images)| 49.97| |BLUEX (No Images) | 43.25| |OAB Exams | 38.13| |Assin2 RTE | 74.75| |Assin2 STS | 71.93| |FaQuAD NLI | 43.97| |HateBR Binary | 57.34| |PT Hate Speech Binary | 60.48| |tweetSentBR | 61.11|