Remove Granite 2B

#1
by adamelliotfields - opened

Should only use models that fit in a single safetensors file; must be under 2B.

As of right now, the only alternative non-thinking model is Gemma 2B, which is the same size as Granite. Thinking models eat up ZeroGPU time, so best saved for API.

adamelliotfields changed discussion title from Replace Granite 2B with Qwen 1.5B R1 Distilled to Remove Granite 2B

Sign up or log in to comment