--- language: - multilingual - fa - en license: cc-by-nc-sa-4.0 library_name: transformers tags: - text-generation-inference - llama-cpp - gguf-my-repo metrics: - bleu - comet - accuracy - perplexity - spearmanr inference: false pipeline_tag: text-generation co2_eq_emissions: emissions: 232380 --- # SmartGitiCorp/PersianMind-v1.0-Q8_0-GGUF This model was converted to GGUF format from [`universitytehran/PersianMind-v1.0`](https://huggingface.co./universitytehran/PersianMind-v1.0) using llama by [`M0javad`](https://huggingface.co./m0javad). Refer to the [original model card](https://huggingface.co./universitytehran/PersianMind-v1.0) for more details on the model. ## Use with llama.cpp Install llama.cpp through brew. ```bash brew install ggerganov/ggerganov/llama.cpp ``` Invoke the llama.cpp server or the CLI. CLI: ```bash llama-cli --hf-repo SmartGitiCorp/PersianMind-v1.0-Q8_0-GGUF --model persianmind-v1.0.Q8_0.gguf -p "معنای زندگی آن است که" ``` Server: ```bash llama-server --hf-repo SmartGitiCorp/PersianMind-v1.0-Q8_0-GGUF --model persianmind-v1.0.Q8_0.gguf -c 2048 ``` Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well. ``` git clone https://github.com/ggerganov/llama.cpp && cd llama.cpp && make && ./main -m persianmind-v1.0.Q8_0.gguf -n 128 ``` @misc {smart_giti_corporation_2024, author = { {Smart Giti Corporation} }, title = { PersianMind-v1.0-Q8_0-GGUF (Revision eb745fe) }, year = 2024, url = { https://huggingface.co./SmartGitiCorp/PersianMind-v1.0-Q8_0-GGUF }, doi = { 10.57967/hf/2364 }, publisher = { Hugging Face } }