DESCRIPTION

Velvet-2B converted to GGUF format (F32) with fbuciuni90/llama.cpp fork and quantized with ggerganov/llama.cpp commit b4689.

NOTE: The Velvet tokenizer is not yet compatible with ggerganov/llama.cpp. Please wait for pull request #11716 to be merged, or compile it yourself.

Original Model: https://huggingface.co./Almawave/Velvet-2B

PROMPT FORMAT

Basic prompt format:

<s><instruction>{prompt}</instruction>

Prompt format with system message:

<s><instruction>{system_prompt}\n\n{prompt}</instruction>

DOWNLOAD

BYE :3

Downloads last month
201
GGUF
Model size
2.22B params
Architecture
llama

3-bit

4-bit

5-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for DagMeow/Velvet-2B-GGUF

Base model

Almawave/Velvet-2B
Quantized
(2)
this model