--- base_model: unsloth/gemma-2-9b language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - llama - trl - sft --- # MedGPT-Gemma2-9B-v.1 - This model is a fine-tuned version of [unsloth/gemma-2-9b](https://huggingface.co./unsloth/gemma-2-9b) on an dataset created by [Valerio Job](https://huggingface.co./valeriojob) together with GPs based on real medical data. - Version 1 (v.1) of MedGPT is the very first version of MedGPT and the training dataset has been kept simple and small with only 60 examples. - This repo includes the 16bit format of the model as well as the LoRA adapters of the model. There is a separate repo called [valeriojob/MedGPT-Gemma2-9B-BA-v.1-GGUF](https://huggingface.co./valeriojob/MedGPT-Gemma2-9B-BA-v.1-GGUF) that includes the quantized versions of this model in GGUF format. - This model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. ## Model description This model acts as a supplementary assistance to GPs helping them in medical and admin tasks. ## Intended uses & limitations The fine-tuned model should not be used in production! This model has been created as a initial prototype in the context of a bachelor thesis. ## Training and evaluation data The dataset (train and test) used for fine-tuning this model can be found here: [datasets/valeriojob/BA-v.1](https://huggingface.co./datasets/valeriojob/BA-v.1) ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - per_device_train_batch_size = 2, - gradient_accumulation_steps = 4, - warmup_steps = 5, - max_steps = 60, - learning_rate = 2e-4, - fp16 = not is_bfloat16_supported(), - bf16 = is_bfloat16_supported(), - logging_steps = 1, - optim = "adamw_8bit", - weight_decay = 0.01, - lr_scheduler_type = "linear", - seed = 3407, - output_dir = "outputs" ### Training results | Training Loss | Step | |:-------------:|:----:| | 2.237900 | 1 | | 2.292200 | 2 | | 2.215200 | 3 | | 1.561200 | 5 | | 0.584500 | 10 | | 0.372500 | 15 | | 0.258600 | 20 | | 0.126300 | 30 | | 0.064100 | 40 | | 0.040800 | 50 | | 0.045700 | 60 | ## Licenses - **License:** apache-2.0