Update README.md
Browse files
README.md
CHANGED
@@ -27,21 +27,25 @@ It uses the Unified AI Description Format (.aidf), which is a novel format with
|
|
27 |
|----------------------------|--------------------------------------------------------------------------------------------------------------|
|
28 |
| **Model** | [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) |
|
29 |
| **Type** | Text Generation |
|
30 |
-
| **
|
|
|
31 |
| **Architecture** | Transformers |
|
32 |
| **File Format** | GGUF |
|
33 |
| **Quantization Types** | 8 Bit <br>5 Bit (K_M) |
|
34 |
-
| **Tools used** | llama.cpp (Commit 9e20231)
|
35 |
| **Model Creator** | [jphme](https://huggingface.co/jphme) |
|
36 |
| **Training Data** | Prorietary German Conversation Dataset, German SQuAD, German legal SQuAD data, augmented with "wrong" contexts, to improve factual RAG. For details see original model link. |
|
37 |
|
38 |
-
**
|
39 |
-
*Profile Type:*
|
40 |
-
*Profile Version:* v1.0
|
41 |
-
|
|
|
42 |
1. Clone and install llama.cpp *(Commit: 9e20231)*.
|
43 |
```
|
44 |
-
# Install llama.cpp by cloning the repo
|
|
|
|
|
45 |
```
|
46 |
2. Use the provided `convert.py` file to convert the original model to GGUF with FP16 precision.
|
47 |
```
|
|
|
27 |
|----------------------------|--------------------------------------------------------------------------------------------------------------|
|
28 |
| **Model** | [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) |
|
29 |
| **Type** | Text Generation |
|
30 |
+
| **Parameters** | 70B |
|
31 |
+
| **Fine-tuned** | Chat, Instruct | |
|
32 |
| **Architecture** | Transformers |
|
33 |
| **File Format** | GGUF |
|
34 |
| **Quantization Types** | 8 Bit <br>5 Bit (K_M) |
|
35 |
+
| **Tools used** | llama.cpp (Commit 9e20231) for quantization to 8, 5 and 4 bit |
|
36 |
| **Model Creator** | [jphme](https://huggingface.co/jphme) |
|
37 |
| **Training Data** | Prorietary German Conversation Dataset, German SQuAD, German legal SQuAD data, augmented with "wrong" contexts, to improve factual RAG. For details see original model link. |
|
38 |
|
39 |
+
**Metadata**
|
40 |
+
*Profile Type:* General AI Profile
|
41 |
+
*Profile Version:* AIDF v1.0
|
42 |
+
|
43 |
+
## Replicate
|
44 |
1. Clone and install llama.cpp *(Commit: 9e20231)*.
|
45 |
```
|
46 |
+
# Install llama.cpp by cloning the repo from Github.
|
47 |
+
# When cloned, then:
|
48 |
+
cd llama.cpp && make
|
49 |
```
|
50 |
2. Use the provided `convert.py` file to convert the original model to GGUF with FP16 precision.
|
51 |
```
|