File size: 1,170 Bytes
eca1bd6 c16e645 eca1bd6 f34b11d eca1bd6 f34b11d eca1bd6 f34b11d 996272b eca1bd6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 |
---
language:
- en
license: apache-2.0
tags:
- text-generation-inference
- transformers
- unsloth
- llama
- trl
- sft
---
# Llama 3.1 Mini - LoRA Finetuned
## Model Description
This model is a LoRA-finetuned version of the Llama 3.1 Mini model, which is a pruned variant of the Llama 3.1 8B model. The original Llama 3.1 Mini was created by pruning the larger model to approximately 3 billion parameters, and this version has been further adapted using Low-Rank Adaptation (LoRA) to enhance its capabilities.
## Limitations
Please note that this model, like its base version, may exhibit biases present in its training data and should be used with appropriate care and consideration.
## Training Data
The base model (Llama 3.1 Mini) was trained on my personal Claude 3 Opus and Claude 3.5 Sonnet dataset, with some synthetic pairs added on with Gemma 2 9B it being the user, and Llama 3 70B through Groq being the assistant. I have also used Guanaco alongside everything else.
## License
Llama 3.1
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|