Safetensors
cohere

Hormoz 8B

Introduction

This model is an effort in order to make a multi-lingual and on device models which can be executed on the consumer hardware. The model follows the steps used in training DeepSeek model. However, the model is not a reasoning model and a generic question answering, conversational and uncensored model which has been made with a cost of around $4000 USD.

If you're curious about the model you also can see our GitHub and learn more about the benchmarks and costs.

Also, this model is based on Command R's architecture, since that architecture gave us the best results in multilingual chat. Specially with languages such as Persian and Arabic. This way, you can consider this model like a commercially useaeble version of aya expanse as well.

The name

The name Hormoz comes from the Persian word "هرمز" which has multiple meanings. It can point to the strait of Hormoz in Persian Gulf or Hormoz Island which is part of the Hormozgan Province in the south of Iran. Also it may point to "اورمزد" or Ourmozd which is middle/ancient Persian name for the planet Jupiter and derived from the term Ahura Mazda or the Avestan term for God.

How to run (transformers)

Install transformers

pip install transformers --upgrade

Note: For better performance, you may need to install accelerate package as well.

Inference

from transformers import AutoTokenizer, AutoModelForCausalLM

model_id = "mann-e/Hormoz-8B"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id).to("cuda")

messages = [{"role": "user", "content": "What is the answer to universe, life and everything?"}]
input_ids = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt").to("cuda")

gen_tokens = model.generate(
    input_ids, 
    max_new_tokens=1024, 
    do_sample=True, 
    temperature=1.0,
    )

gen_text = tokenizer.decode(gen_tokens[0])
print(gen_text)

License

This model is published under MIT license.

Commercial Use

Since this model is MIT licensed, you're free to do whatever you want with the model. However since we're a relatively small startup, we recommend you if you are a big corporate and you host this model, give us a capacity of your API as well. This way, we both can benefit from the model.

Downloads last month
11
Safetensors
Model size
8.03B params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Model tree for mann-e/Hormoz-8B

Quantizations
1 model

Datasets used to train mann-e/Hormoz-8B