GGUF
llama.cpp
Inference Endpoints

Is there any hope to make 500M and 1B parameters like apple OpenELM?

#4
by yousef1727 - opened

I'm just wondering what if there is these sizes? that would be nice special on Android device

Some apps need tiny, mini or medium models for simple tasks not for QnA, i mean summarize, explain, text editing, etc

I so trust on google and I know maybe I'll see more models on future, Google Gemma is the only model who understands Arabic because Gemma Tokenizer split word not letters like gpt which mean he can understand meaning between words not letters, that reduce mistakes,

Finally I want to repeat Apple made big family of sizes why google don't?

Sign up or log in to comment