twosmoothslateslabs commited on
Commit
42a3c6e
·
verified ·
1 Parent(s): 05ec022

Upload 2 files

Browse files
Files changed (2) hide show
  1. README.md +45 -0
  2. measurement.json +0 -0
README.md ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ language:
4
+ - en
5
+ - fr
6
+ - it
7
+ - pt
8
+ - hi
9
+ - es
10
+ - th
11
+ - de
12
+ base_model:
13
+ - ockerman0/MN-12B-Starcannon-v5.5-unofficial
14
+ base_model_relation: quantized
15
+ tags:
16
+ - mergekit
17
+ - merge
18
+ - mistral
19
+ quantized_by: twosmoothslateslabs
20
+ pipeline_tag: text-generation
21
+ ---
22
+
23
+ ## This repo contains EXL2 quants of ockerman0/MN-12B-Starcannon-v5.5-unofficial.
24
+
25
+ Find the original model card [here](https://huggingface.co/ockerman0/MN-12B-Starcannon-v5.5-unofficial).
26
+
27
+ ## Base repo only contains the measurement file, see revisions for the quants.
28
+
29
+ - [measurement.json](https://huggingface.co/anthracite-org/magnum-v4-123b-exl2/tree/main)
30
+ - [3.0bpw](https://huggingface.co/anthracite-org/magnum-v4-123b-exl2/tree/3.0bpw)
31
+ - [4.0bpw](https://huggingface.co/anthracite-org/magnum-v4-123b-exl2/tree/4.0bpw)
32
+ - [5.0bpw](https://huggingface.co/anthracite-org/magnum-v4-123b-exl2/tree/5.0bpw)
33
+ - [6.0bpw](https://huggingface.co/anthracite-org/magnum-v4-123b-exl2/tree/6.0bpw)
34
+
35
+ ## Notes
36
+ Making these was a lesson in pain and humility. It has been over two months since the day I decided "hm today i will learn how to make exl2 quants" <- (clueless). First my conda env
37
+ stopped working (for some reason), then it stopped recognizing venvs when I tried using those, then the universe decided to screw up the one venv I had working somehow (I can only assume
38
+ it was a cosmic bitflip or something because it literally stopped working overnight) and making these four quants alone took over an hour on my hardware, in which time I could probably have
39
+ made an entire set of GGUFs (plus a full set of i-quants) for three different models. Then uploading these was such a pain because huggingface-cli might as well be arcane magic since their
40
+ documentation doesn't really tell you how to actually use it or what exactly will happen when you run this method. I haven't even tested any of these quants because tabbyapi, to this day,
41
+ simply will not work. Torch keeps bugging me about running out of VRAM, even when trying to load 3bs. I have basically tried everything to try and get tabbyapi to run. It simply will not.
42
+
43
+ Suggest more models in the community tab and I might have a crack at exl2'ing them.
44
+
45
+ The model is quite nice though, it is quite useful for my usecases of synthetic variant generation if you're into that sort of thing.
measurement.json ADDED
The diff for this file is too large to render. See raw diff