base_model:
- Bllossom/llama-3-Korean-Bllossom-70B
library_name: transformers
tags:
- mergekit
- merge
ππ Huggingface Upload Issue
Maximum individual file size is 50.0GB to upload huggingface. To clear it, Split the file into part_aa, part_ab, part_ac chunks. After you download this repo on your folder path, command like this
cat part_* > llama-3-korean-bllossom-120b-Q4_K_M.gguf
Perhaps this will be the first GGUF model to upload such a large GGUF file of over 50GB to huggingface?
Sometimes we need to use a trick called chunks.
Please wait to upload.....
π°π· About the JayLee "AsianSoul"
"A leader who can make you rich π΅ !!!"
"Prove yourself with actual results, not just saying I know more than you!!!"
About this model
This is a 128B model based on Bllossom/llama-3-Korean-Bllossom-70B
β I started this Korean 120B model merge while drinking an iced Americano at Starbucks.
If you walk around Starbucks in Seoul, Korea, you will see someone creating a merge and an application based on it.
At that time, please come up to me and say "hello".
ποΈ I am a person whose goal is to turn the great results created by great genius scientists & groups around the world into profitable ones.
My role model is J. Robert Oppenheimer!!!
J. Robert Oppenheimer is highly regarded for his ability to gather and lead a team of brilliant scientists, merging their diverse expertise and efforts towards a common goal.
[Learn more about J. Robert Oppenheimer](https://en.wikipedia.org/wiki/J._Robert_Oppenheimer).
I hope this 120B is a helpful model for your future.
π Collaboration is always welcome π
π You can't beat these giant corporations & groups alone and you can never become rich.
Now we have to come together.
People who can actually become rich together, let's collaborate with me.!!! πΈ
About Bllossom/llama-3-Korean-Bllossom-70B
- Full model released in Korean over 100GB by Blossom team
- First in Korean! Expansion of Korean vocabulary to over 30,000 words
- Capable of processing Korean context that is approximately 25% longer than Llama3
- Connecting Korean-English knowledge using the Korean-English Parallel Corpus (pre-study)
- Fine tuning using data produced by linguists considering Korean culture and language
- Reinforcement learning
π°οΈ About asiansoul/llama-3-Korean-Bllossom-120B-GGUF
- Q4_K_M : On a GPU with 68GB / more OR a CPU with 68G / more memory
- More Quantization ones i hope to upload, but your computer won't be able to handle it then. you know what i mean!!
Models Merged
The following models were included in the merge:
Ollama
Check the information indicated above and run it when your computer is ready.
Otherwise, your computer will freeze.
- Create
ollama create Bllossom -f ./Modelfile_Q4_K_M
- MODELFILE (you can change this for your preference)
FROM llama-3-korean-bllossom-120b-Q4_K_M.gguf
TEMPLATE """{{ if .System }}<|start_header_id|>system<|end_header_id|>
{{ .System }}<|eot_id|>{{ end }}{{ if .Prompt }}<|start_header_id|>user<|end_header_id|>
{{ .Prompt }}<|eot_id|>{{ end }}<|start_header_id|>assistant<|end_header_id|>
{{ .Response }}<|eot_id|>"""
SYSTEM """
μΉμ ν μ±λ΄μΌλ‘μ μλλ°©μ μμ²μ μ΅λν μμΈνκ³ μΉμ νκ² λ΅νμ. λͺ¨λ λλ΅μ νκ΅μ΄(Korean)μΌλ‘ λλ΅ν΄μ€.
"""
PARAMETER num_keep 24
PARAMETER temperature 0.7
PARAMETER num_predict 3000
PARAMETER stop "<|start_header_id|>"
PARAMETER stop "<|end_header_id|>"
PARAMETER stop "<|eot_id|>"
Configuration
The following YAML configuration was used to produce this model:
slices:
- sources:
- layer_range: [0, 20]
model: Bllossom/llama-3-Korean-Bllossom-70B
- sources:
- layer_range: [10, 30]
model: Bllossom/llama-3-Korean-Bllossom-70B
- sources:
- layer_range: [20, 40]
model: Bllossom/llama-3-Korean-Bllossom-70B
- sources:
- layer_range: [30, 50]
model: Bllossom/llama-3-Korean-Bllossom-70B
- sources:
- layer_range: [40, 60]
model: Bllossom/llama-3-Korean-Bllossom-70B
- sources:
- layer_range: [50, 70]
model: Bllossom/llama-3-Korean-Bllossom-70B
- sources:
- layer_range: [60, 80]
model: Bllossom/llama-3-Korean-Bllossom-70B
merge_method: passthrough
dtype: float16