|
--- |
|
libray_name: transformers |
|
pipeline_tag: text-generation |
|
license: other |
|
license_name: llama3 |
|
license_link: LICENSE |
|
language: |
|
- en |
|
- ko |
|
tags: |
|
- meta |
|
- llama |
|
- llama-3 |
|
- akallama |
|
library_name: transformers |
|
--- |
|
<a href="https://github.com/"> |
|
<img src="3de500aklm" width="50%"/> |
|
</a> |
|
|
|
|
|
# AKALLAMA |
|
We introduce AKALLAMA-70B, korean focused opensource 70b large language model. |
|
It demonstrate considerable improvement in korean fluence, specially compared to base llama 3 model. |
|
To our knowledge, this is one of the first 70b opensource Korean-speaking language models. |
|
|
|
### Model Description |
|
|
|
This is the model card of a 🤗 transformers model that has been pushed on the Hub. |
|
|
|
- **Developed by:** [mirlab](https://mirlab.yonsei.ac.kr/) |
|
- **Language(s) (NLP):** Korean, English |
|
- **License:** llama3 |
|
- **Finetuned from model:** [meta-llama/Meta-Llama-3-70B](https://huggingface.co./meta-llama/Meta-Llama-3-70B) |
|
|
|
## Evaluation |
|
|
|
For local inferencing and evaluation, we highly recommend using the Ollama library. |
|
Check _Customize a model section_ of [Ollama github page](https://github.com/ollama/ollama) |
|
|
|
## Training Details |
|
### Training Procedure |
|
|
|
We closely followed training procedure of Zephyr ORPO model. |
|
Please check out Huggingface's [alignment handbook](https://github.com/huggingface/alignment-handbook?tab=readme-ov-file) for further details, including the chat template. |
|
|
|
### Training Data |
|
|
|
Detailed descriptions regarding training data will be announced later. |
|
|
|
### Examples |
|
|
|
## Thanks to |
|
|
|
- A100 클러스터를 제공해주신, 연세대학교 인공지능학과 데이터센터 |
|
- |
|
|
|
|