|
--- |
|
language: |
|
- en |
|
pipeline_tag: text-generation |
|
license: cc-by-nc-sa-4.0 |
|
datasets: |
|
- nlpai-lab/databricks-dolly-15k-ko |
|
- kyujinpy/KOR-OpenOrca-Platypus-v3 |
|
--- |
|
|
|
**Input** Models input text only. |
|
|
|
**Output** Models generate text only. |
|
|
|
**Base Model** [beomi/Yi-Ko-6B](https://huggingface.co./beomi/Yi-Ko-6B) |
|
|
|
**Training Dataset** |
|
- [nlpai-lab/databricks-dolly-15k-ko](https://huggingface.co./datasets/nlpai-lab/databricks-dolly-15k-ko) |
|
- [kyujinpy/KOR-OpenOrca-Platypus-v3](https://huggingface.co./datasets/kyujinpy/KOR-OpenOrca-Platypus-v3) |
|
|
|
# Implementation Code |
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
import torch |
|
repo = "ifuseok/yi-ko-playtus-instruct-v0.2" |
|
OpenOrca = AutoModelForCausalLM.from_pretrained( |
|
repo, |
|
return_dict=True, |
|
torch_dtype=torch.float16, |
|
device_map='auto' |
|
) |
|
OpenOrca_tokenizer = AutoTokenizer.from_pretrained(repo) |
|
``` |
|
|
|
# Prompt Example |
|
``` |
|
<|system|> |
|
μμ€ν
λ©μμ§ μ
λλ€. <|endoftext|> |
|
<|user|> |
|
μ μ μ
λλ€.<|endoftext|> |
|
<|assistant|> |
|
μ΄μμ€ν΄νΈ μ
λλ€.<|endoftext|> |
|
``` |