eachadea commited on
Commit
17ebbf7
1 Parent(s): df20cac

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -10
README.md CHANGED
@@ -4,19 +4,11 @@ tags:
4
  - vicuna
5
  - llama
6
  - text-generation-inference
7
- ---
8
- Converted for use with [llama.cpp](https://github.com/ggerganov/llama.cpp)
9
- ---
10
- - 4-bit quantized
11
- - Needs ~10GB of CPU RAM
12
- - Won't work with alpaca.cpp or old llama.cpp (new ggml format requires latest llama.cpp)
13
- - EOS token fix added (download rev1)
14
-
15
  ---
16
 
17
- If you only have 8GB RAM, a smaller 7B version of this can be found here: https://huggingface.co/eachadea/ggml-vicuna-7b-4bit.
 
18
 
19
- 7B is over 2x faster and is also uncensored, while 13B isn't.
20
 
21
  ---
22
  tags:
 
4
  - vicuna
5
  - llama
6
  - text-generation-inference
 
 
 
 
 
 
 
 
7
  ---
8
 
9
+ **NOTE: Download new version here: https://huggingface.co/eachadea/ggml-vicuna-13b-1.1**
10
+ **NOTE: Download new version (7B): https://huggingface.co/eachadea/ggml-vicuna-7b-1.1**
11
 
 
12
 
13
  ---
14
  tags: