|
--- |
|
license: cc-by-nc-4.0 |
|
datasets: |
|
- kyujinpy/KOR-OpenOrca-Platypus-v3 |
|
- beomi/KoAlpaca-v1.1a |
|
- maywell/ko_wikidata_QA |
|
language: |
|
- ko |
|
base_model: beomi/Yi-Ko-6B |
|
--- |
|
|
|
# Yi-Ko-6B-Instruct-v1.1 |
|
|
|
## Model Details |
|
|
|
### Base Model |
|
[beomi/Yi-Ko-6B](https://huggingface.co./beomi/Yi-Ko-6B) |
|
|
|
### Training Dataset |
|
1. [kyujinpy/KOR-OpenOrca-Platypus-v3](https://huggingface.co./datasets/kyujinpy/KOR-OpenOrca-Platypus-v3) π |
|
2. [beomi/KoAlpaca-v1.1a](https://huggingface.co./datasets/beomi/KoAlpaca-v1.1a) π |
|
3. [maywell/ko_wikidata_QA](https://huggingface.co./datasets/maywell/ko_wikidata_QA) π |
|
4. AIHub λ°μ΄ν° νμ© |
|
|
|
## Instruction Format |
|
```python |
|
### User: |
|
{instruction} |
|
|
|
### Assistant: |
|
{response} |
|
``` |
|
|
|
## Loading the Model |
|
```python |
|
import torch |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
tokenizer = AutoTokenizer.from_pretrained("wkshin89/Yi-Ko-6B-Instruct-v1.1") |
|
model = AutoModelForCausalLM.from_pretrained( |
|
"wkshin89/Yi-Ko-6B-Instruct-v1.1", |
|
device_map="auto", |
|
torch_dtype=torch.bfloat16, |
|
) |
|
``` |