|
--- |
|
base_model: yanolja/EEVE-Korean-Instruct-10.8B-v1.0 |
|
license: apache-2.0 |
|
language: |
|
- ko |
|
- en |
|
--- |
|
|
|
# **Jolteon-Instruct-13B-alpha** |
|
|
|
The model was trained based on the [EEVE-Korean-Instruct-10.8B-v1.0](https://huggingface.co./yanolja/EEVE-Korean-Instruct-10.8B-v1.0) model from [yanolja](https://www.yanolja.com), extended to 13.4b (12 layer pass-through) utilizing [mergekit](https://github.com/cg123/mergekit). |
|
|
|
## Methodology |
|
|
|
TBD |
|
|
|
## Training Details |
|
| |Training Data|Parameters|Content Length|Samples Seen|Learning Rate| |
|
|---|---|---|---|---|---| |
|
|Jolteon-Instruct-13B-alpha|*A curated mix of English + Korean Instruction set*|13.4B|4k|>400k|1e<sup>-5</sup>| |
|
|
|
## Example |
|
|
|
## License |
|
|
|
본 모델은 apache-2.0 라이센스를 따릅니다. 모델을 사용하여 생성된 데이터셋을 배포할 경우 모델 사용을 명시해 주시기를 권고드립니다. |
|
|
|
## Thanks to |
|
|
|
- A100 클러스터를 제공해주신, [Sionic AI](https://sionic.ai/) |
|
|
|
## Contact |
|
- [Discord Server Link](https://discord.gg/MrBt3PXdXc) |