Upload Mistral-NeMo-Minitron-8B-Chat.Q8_0.gguf with huggingface_hub
Browse files
.gitattributes
CHANGED
@@ -36,3 +36,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
36 |
Mistral-NeMo-Minitron-8B-Chat.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
Mistral-NeMo-Minitron-8B-Chat.Q4_1.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
Mistral-NeMo-Minitron-8B-Chat.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
36 |
Mistral-NeMo-Minitron-8B-Chat.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
Mistral-NeMo-Minitron-8B-Chat.Q4_1.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
Mistral-NeMo-Minitron-8B-Chat.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
|
39 |
+
Mistral-NeMo-Minitron-8B-Chat.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
Mistral-NeMo-Minitron-8B-Chat.Q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d0777fcb31a60811945abba0f77ab5c986f1c5578c344f8617a64ad8f4b68fc3
|
3 |
+
size 8948861344
|