jeevan-exa commited on
Commit
eb6b60d
·
verified ·
1 Parent(s): cd6969e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +70 -2
README.md CHANGED
@@ -15,6 +15,76 @@ tags:
15
  - Open-MedQA-Nexus
16
  ---
17
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  # Uploaded model
19
 
20
  - **Developed by:** exafluence
@@ -22,5 +92,3 @@ tags:
22
  - **Finetuned from model :** unsloth/mistral-nemo-instruct-2407-bnb-4bit
23
 
24
  This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
25
-
26
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
15
  - Open-MedQA-Nexus
16
  ---
17
 
18
+ # EXF-Medistral-Nemo-12B
19
+
20
+ ## Model Description
21
+ **EXF-Medistral-Nemo-12B** is a fine-tuned version of the **Mistral-Nemo-12B** model, optimized for tasks in the medical domain. It has been trained on the **Open-Nexus-MedQA** dataset, which integrates a wide range of medical knowledge from public datasets like **ChatDoctor**, **icliniq**, and others, to enhance the model’s ability to answer medical questions accurately and reliably. This model is designed to assist in clinical decision support, medical coding, and patient care by generating responses based on comprehensive medical knowledge.
22
+
23
+ ## Model Architecture
24
+ - **Base Model**: Mistral-Nemo-12B
25
+ - **Parameters**: 12 billion
26
+ - **Fine-tuning Dataset**: Open-Nexus-MedQA
27
+ - **Task**: Medical question-answering (QA), medical coding, and healthcare information retrieval.
28
+
29
+ ## Training Data
30
+ The model was fine-tuned on the **Open-Nexus-MedQA** dataset, which aggregates data from multiple medical QA sources such as:
31
+ - **ChatDoctor**
32
+ - **icliniq.com**
33
+ - **HealthCareMagic**
34
+ - **CareQA**
35
+ - **MedInstruct**
36
+
37
+ The dataset contains medical queries ranging from simple conditions to complex diagnoses, accompanied by accurate, domain-specific responses, making it a robust training source for real-world medical applications.
38
+
39
+ ## Intended Use
40
+ **EXF-Medistral-Nemo-12B** is ideal for:
41
+ - **Medical Question-Answering**: It can be used for generating responses to patient queries or supporting healthcare professionals with clinical information.
42
+ - **Medical Coding**: The model supports tasks related to **CMS**, **OASIS**, **ICD-10**, and other coding systems.
43
+ - **Clinical Decision Support**: Assisting doctors and healthcare providers by offering evidence-based suggestions or answers.
44
+ - **Patient Care Tools**: Powering medical chatbots or virtual assistants for patients seeking health information.
45
+
46
+ ## Performance
47
+ The model has been fine-tuned for precision in the medical domain, demonstrating high accuracy in understanding and generating responses to complex medical queries. It excels in:
48
+ - **Medical terminology comprehension**
49
+ - **Providing accurate ICD-10 and CMS codes**
50
+ - **Generating medically relevant and safe answers**
51
+
52
+ ## Limitations
53
+ - **Not a Diagnostic Tool**: This model is not intended to replace medical professionals or provide definitive medical diagnoses. Always consult with a licensed healthcare provider for medical advice.
54
+ - **Training Data Bias**: The dataset is based on publicly available medical QA data, which might not cover all edge cases or international healthcare systems.
55
+
56
+ ## How to Use
57
+
58
+ ```python
59
+ from transformers import AutoModelForCausalLM, AutoTokenizer
60
+ tokenizer = AutoTokenizer.from_pretrained("exafluence/EXF-Medistral-Nemo-12B-4bit")
61
+ model = AutoModelForCausalLM.from_pretrained("exafluence/EXF-Medistral-Nemo-12B-4bit")
62
+ input_text = "What are the symptoms of type 2 diabetes?"
63
+ inputs = tokenizer(input_text, return_tensors="pt")
64
+ outputs = model.generate(**inputs)
65
+ print(tokenizer.decode(outputs, skip_special_tokens=True))
66
+ ```
67
+
68
+ ## License
69
+
70
+ This model is provided under a proprietary license. Usage is restricted to non-commercial purposes unless explicit permission is granted.
71
+
72
+ Citation
73
+ If you use this model, please cite:
74
+ ```bibtex
75
+ @inproceedings{exafluence2024EXFMedistralNemo12B,
76
+ title={EXF-Medistral-Nemo-12B: A Fine-Tuned Medical Language Model for Healthcare Applications},
77
+ author={Exafluence Inc.},
78
+ year={2024},
79
+ url={https://huggingface.co/exafluence/EXF-Medistral-Nemo-12B}
80
+ doi={https://doi.org/10.57967/hf/3284}
81
+ }
82
+ ```
83
+
84
+ ## Contact
85
+
86
+ For any questions or inquiries regarding usage, licensing, or access, please contact Exafluence Inc.
87
+
88
  # Uploaded model
89
 
90
  - **Developed by:** exafluence
 
92
  - **Finetuned from model :** unsloth/mistral-nemo-instruct-2407-bnb-4bit
93
 
94
  This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.