Synatra-7B-v0.3-base๐ง
Support Me
์๋ํธ๋ผ๋ ๊ฐ์ธ ํ๋ก์ ํธ๋ก, 1์ธ์ ์์์ผ๋ก ๊ฐ๋ฐ๋๊ณ ์์ต๋๋ค. ๋ชจ๋ธ์ด ๋ง์์ ๋์ จ๋ค๋ฉด ์ฝ๊ฐ์ ์ฐ๊ตฌ๋น ์ง์์ ์ด๋จ๊น์?
Wanna be a sponser? Contact me on Telegram AlzarTakkarsen
License
This model is strictly non-commercial (cc-by-nc-4.0) use only. The "Model" is completely free (ie. base model, derivates, merges/mixes) to use for non-commercial purposes as long as the the included cc-by-nc-4.0 license in any parent repository, and the non-commercial use statute remains, regardless of other models' licences. The licence can be changed after new model released. If you are to use this model for commercial purpose, Contact me.
Model Details
Base Model
mistralai/Mistral-7B-Instruct-v0.1
Trained On
A6000 48GB * 8
Instruction format
It follows ChatML format and Alpaca(No-Input) format.
TODO
โRP ๊ธฐ๋ฐ ํ๋ ๋ชจ๋ธ ์ ์
โ๋ฐ์ดํฐ์ ์ ์
- ์ธ์ด ์ดํด๋ฅ๋ ฅ ๊ฐ์
โ์์ ๋ณด์
- ํ ํฌ๋์ด์ ๋ณ๊ฒฝ
Model Benchmark
Ko-LLM-Leaderboard
On Benchmarking...
Implementation Code
Since, chat_template already contains insturction format above. You can use the code below.
from transformers import AutoModelForCausalLM, AutoTokenizer
device = "cuda" # the device to load the model onto
model = AutoModelForCausalLM.from_pretrained("maywell/Synatra-7B-v0.3-base")
tokenizer = AutoTokenizer.from_pretrained("maywell/Synatra-7B-v0.3-base")
messages = [
{"role": "user", "content": "๋ฐ๋๋๋ ์๋ ํ์์์ด์ผ?"},
]
encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
model_inputs = encodeds.to(device)
model.to(device)
generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
decoded = tokenizer.batch_decode(generated_ids)
print(decoded[0])
- Downloads last month
- 4,919