--- base_model: - Bllossom/llama-3-Korean-Bllossom-70B library_name: transformers tags: - mergekit - merge --- πŸŒ‹πŸŒ‹ Huggingface Upload Issue Maximum individual file size is 50.0GB to upload huggingface. To clear it, Split the file into part_aa, part_ab, part_ac chunks. After you download this repo on your folder path, command like this ``` cat part_* > llama-3-korean-bllossom-120b-Q4_K_M.gguf ``` Please wait to upload..... ### πŸ‡°πŸ‡· About the JayLee "AsianSoul" ``` "A leader who can make you rich !!!" "Prove yourself with actual results, not just saying I know more than you!!!" ``` Screenshot-2024-05-18-at-11-08-12-PM ### About this model This is a 128B model based on [Bllossom/llama-3-Korean-Bllossom-70B](https://huggingface.co./Bllossom/llama-3-Korean-Bllossom-70B) To know more about the base model, check link. I hope this 120B is a helpful model for your future. 🌍 Collaboration is always welcome.🌍 ``` About Bllossom/llama-3-Korean-Bllossom-70B - Full model released in Korean over 100GB by Blossom team - First in Korean! Expansion of Korean vocabulary to over 30,000 words - Capable of processing Korean context that is approximately 25% longer than Llama3 - Connecting Korean-English knowledge using the Korean-English Parallel Corpus (pre-study) - Fine tuning using data produced by linguists considering Korean culture and language - Reinforcement learning πŸ›°οΈ About asiansoul/llama-3-Korean-Bllossom-120B-GGUF - Q4_K_M : On a GPU with 68GB / more OR a CPU with 68G / more memory ``` ### Models Merged The following models were included in the merge: * [Bllossom/llama-3-Korean-Bllossom-70B](https://huggingface.co./Bllossom/llama-3-Korean-Bllossom-70B) ### Ollama Create ``` ollama create Bllossom -f ./Modelfile_Q4_K_M ``` MODELFILE (you can change this for your preference) ``` FROM llama-3-korean-bllossom-120b-Q4_K_M.gguf TEMPLATE """{{ if .System }}<|start_header_id|>system<|end_header_id|> {{ .System }}<|eot_id|>{{ end }}{{ if .Prompt }}<|start_header_id|>user<|end_header_id|> {{ .Prompt }}<|eot_id|>{{ end }}<|start_header_id|>assistant<|end_header_id|> {{ .Response }}<|eot_id|>""" SYSTEM """ μΉœμ ˆν•œ μ±—λ΄‡μœΌλ‘œμ„œ μƒλŒ€λ°©μ˜ μš”μ²­μ— μ΅œλŒ€ν•œ μžμ„Έν•˜κ³  μΉœμ ˆν•˜κ²Œ λ‹΅ν•˜μž. λͺ¨λ“  λŒ€λ‹΅μ€ ν•œκ΅­μ–΄(Korean)으둜 λŒ€λ‹΅ν•΄μ€˜. """ PARAMETER num_keep 24 PARAMETER temperature 0.7 PARAMETER num_predict 3000 PARAMETER stop "<|start_header_id|>" PARAMETER stop "<|end_header_id|>" PARAMETER stop "<|eot_id|>" ``` ### Configuration The following YAML configuration was used to produce this model: ```yaml slices: - sources: - layer_range: [0, 20] model: Bllossom/llama-3-Korean-Bllossom-70B - sources: - layer_range: [10, 30] model: Bllossom/llama-3-Korean-Bllossom-70B - sources: - layer_range: [20, 40] model: Bllossom/llama-3-Korean-Bllossom-70B - sources: - layer_range: [30, 50] model: Bllossom/llama-3-Korean-Bllossom-70B - sources: - layer_range: [40, 60] model: Bllossom/llama-3-Korean-Bllossom-70B - sources: - layer_range: [50, 70] model: Bllossom/llama-3-Korean-Bllossom-70B - sources: - layer_range: [60, 80] model: Bllossom/llama-3-Korean-Bllossom-70B merge_method: passthrough dtype: float16 ```