|
--- |
|
datasets: |
|
- beomi/KoAlpaca-v1.1a |
|
language: |
|
- ko |
|
- en |
|
base_model: |
|
- ibm-granite/granite-3.1-8b-instruct |
|
library_name: transformers |
|
--- |
|
|
|
## Model Details |
|
|
|
**Granite-3.1-8B-instruct-KR** |
|
|
|
Granite-3.1-8B-instruct-KR is continued pretrained(fully fine-tuned) language model based on Granite-3.1-8B-Instruct. |
|
|
|
This model is trained fully with publicily available resource at HuggingFace dataset hub, preprocessed Korean texts. |
|
|
|
The train was done on A6000 48GB * 4. |
|
|
|
**Model developers** Dongwook Min (mindw96) |
|
|
|
**Dataset** beomi/KoAlpaca-v1.1a |
|
|
|
**Variations** Granite-3.1-8B-instruct-KR comes in one size — 8B. |
|
|
|
**Input** Models input text only. |
|
|
|
**Output** Models generate text only. |
|
|
|
**Model Architecture** Granite 3.1 is an auto-regressive language model that uses an optimized transformer architecture. |
|
|
|
**Model Release Date** 02.01.2025. |