freefallr commited on
Commit
605629f
1 Parent(s): ea18b32

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -7
README.md CHANGED
@@ -27,21 +27,25 @@ It uses the Unified AI Description Format (.aidf), which is a novel format with
27
  |----------------------------|--------------------------------------------------------------------------------------------------------------|
28
  | **Model** | [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) |
29
  | **Type** | Text Generation |
30
- | **Finetuned for** | Chat | |
 
31
  | **Architecture** | Transformers |
32
  | **File Format** | GGUF |
33
  | **Quantization Types** | 8 Bit <br>5 Bit (K_M) |
34
- | **Tools used** | llama.cpp (Commit 9e20231) |
35
  | **Model Creator** | [jphme](https://huggingface.co/jphme) |
36
  | **Training Data** | Prorietary German Conversation Dataset, German SQuAD, German legal SQuAD data, augmented with "wrong" contexts, to improve factual RAG. For details see original model link. |
37
 
38
- **Profile Metadata**
39
- *Profile Type:*
40
- *Profile Version:* v1.0
41
- ## Replication Steps
 
42
  1. Clone and install llama.cpp *(Commit: 9e20231)*.
43
  ```
44
- # Install llama.cpp by cloning the repo and compiling it.
 
 
45
  ```
46
  2. Use the provided `convert.py` file to convert the original model to GGUF with FP16 precision.
47
  ```
 
27
  |----------------------------|--------------------------------------------------------------------------------------------------------------|
28
  | **Model** | [jphme/Llama-2-13b-chat-german](https://huggingface.co/jphme/Llama-2-13b-chat-german) |
29
  | **Type** | Text Generation |
30
+ | **Parameters** | 70B |
31
+ | **Fine-tuned** | Chat, Instruct | |
32
  | **Architecture** | Transformers |
33
  | **File Format** | GGUF |
34
  | **Quantization Types** | 8 Bit <br>5 Bit (K_M) |
35
+ | **Tools used** | llama.cpp (Commit 9e20231) for quantization to 8, 5 and 4 bit |
36
  | **Model Creator** | [jphme](https://huggingface.co/jphme) |
37
  | **Training Data** | Prorietary German Conversation Dataset, German SQuAD, German legal SQuAD data, augmented with "wrong" contexts, to improve factual RAG. For details see original model link. |
38
 
39
+ **Metadata**
40
+ *Profile Type:* General AI Profile
41
+ *Profile Version:* AIDF v1.0
42
+
43
+ ## Replicate
44
  1. Clone and install llama.cpp *(Commit: 9e20231)*.
45
  ```
46
+ # Install llama.cpp by cloning the repo from Github.
47
+ # When cloned, then:
48
+ cd llama.cpp && make
49
  ```
50
  2. Use the provided `convert.py` file to convert the original model to GGUF with FP16 precision.
51
  ```