Triangle104 commited on
Commit
68b2919
·
verified ·
1 Parent(s): f2e9ab6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +118 -1
README.md CHANGED
@@ -4,12 +4,129 @@ base_model: P0x0/Epos-8b
4
  tags:
5
  - llama-cpp
6
  - gguf-my-repo
 
7
  ---
8
 
9
  # Triangle104/Epos-8b-Q4_K_M-GGUF
10
  This model was converted to GGUF format from [`P0x0/Epos-8b`](https://huggingface.co/P0x0/Epos-8b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
11
  Refer to the [original model card](https://huggingface.co/P0x0/Epos-8b) for more details on the model.
12
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  ## Use with llama.cpp
14
  Install llama.cpp through brew (works on Mac and Linux)
15
 
@@ -48,4 +165,4 @@ Step 3: Run inference through the main binary.
48
  or
49
  ```
50
  ./llama-server --hf-repo Triangle104/Epos-8b-Q4_K_M-GGUF --hf-file epos-8b-q4_k_m.gguf -c 2048
51
- ```
 
4
  tags:
5
  - llama-cpp
6
  - gguf-my-repo
7
+ license: llama3.1
8
  ---
9
 
10
  # Triangle104/Epos-8b-Q4_K_M-GGUF
11
  This model was converted to GGUF format from [`P0x0/Epos-8b`](https://huggingface.co/P0x0/Epos-8b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
12
  Refer to the [original model card](https://huggingface.co/P0x0/Epos-8b) for more details on the model.
13
 
14
+ ---
15
+ Model details:
16
+ -
17
+ Epos-8B is a fine-tuned version of the base model Llama-3.1-8B
18
+ from Meta, optimized for storytelling, dialogue generation, and
19
+ creative writing. The model specializes in generating rich narratives,
20
+ immersive prose, and dynamic character interactions, making it ideal for
21
+ creative tasks.
22
+
23
+ Model Description
24
+
25
+
26
+
27
+
28
+ Epos-8B is an 8 billion parameter language model fine-tuned for
29
+ storytelling and narrative tasks. Inspired by the grandeur of epic
30
+ tales, it is designed to produce high-quality, engaging content that
31
+ evokes the depth and imagination of ancient myths and modern
32
+ storytelling traditions.
33
+
34
+
35
+ Developed by: P0x0
36
+ Funded by: P0x0
37
+ Shared by: P0x0
38
+ Model type: Transformer-based Language Model
39
+ Language(s) (NLP): Primarily English
40
+ License: Apache 2.0
41
+ Finetuned from model: meta-llama/Llama-3.1-8B
42
+
43
+
44
+
45
+
46
+
47
+
48
+
49
+ Model Sources
50
+
51
+
52
+
53
+
54
+ Repository: Epos-8B on Hugging Face
55
+ GGUF Repository: Epos-8B-GGUF (TO BE ADDED)
56
+
57
+
58
+
59
+
60
+
61
+
62
+
63
+
64
+ Uses
65
+
66
+
67
+
68
+
69
+
70
+
71
+
72
+
73
+
74
+ Direct Use
75
+
76
+
77
+
78
+
79
+ Epos-8B is ideal for:
80
+
81
+
82
+ Storytelling: Generate detailed, immersive, and engaging narratives.
83
+ Dialogue Creation: Create realistic and dynamic character interactions for stories or games.
84
+
85
+
86
+
87
+
88
+
89
+
90
+
91
+ How to Get Started with the Model
92
+
93
+
94
+
95
+
96
+ To run the quantized version of the model, you can use KoboldCPP, which allows you to run quantized GGUF models locally.
97
+
98
+
99
+
100
+
101
+
102
+
103
+
104
+ Steps:
105
+
106
+
107
+
108
+
109
+ Download KoboldCPP.
110
+ Follow the setup instructions provided in the repository.
111
+ Download the GGUF variant of Epos-8B from Epos-8B-GGUF.
112
+ Load the model in KoboldCPP and start generating!
113
+
114
+
115
+ Alternatively, integrate the model directly into your code with the following snippet:
116
+
117
+
118
+ from transformers import AutoModelForCausalLM, AutoTokenizer
119
+
120
+ tokenizer = AutoTokenizer.from_pretrained("P0x0/Epos-8B")
121
+ model = AutoModelForCausalLM.from_pretrained("P0x0/Epos-8B")
122
+
123
+ input_text = "Once upon a time in a distant land..."
124
+ inputs = tokenizer(input_text, return_tensors="pt")
125
+ outputs = model.generate(**inputs)
126
+
127
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
128
+
129
+ ---
130
  ## Use with llama.cpp
131
  Install llama.cpp through brew (works on Mac and Linux)
132
 
 
165
  or
166
  ```
167
  ./llama-server --hf-repo Triangle104/Epos-8b-Q4_K_M-GGUF --hf-file epos-8b-q4_k_m.gguf -c 2048
168
+ ```