DavidAU commited on
Commit
adbfece
·
verified ·
1 Parent(s): 9ab901e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +27 -0
README.md CHANGED
@@ -146,6 +146,33 @@ Longer prompts vastly increase the quality of the model's output (see later exam
146
 
147
  You may want to use "regular" Dark Planet 8B [https://huggingface.co/DavidAU/L3-Dark-Planet-8B-GGUF] for some writing task(s), and this model for prose specific task(s).
148
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
149
  <B>Model Template:</B>
150
 
151
  This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
 
146
 
147
  You may want to use "regular" Dark Planet 8B [https://huggingface.co/DavidAU/L3-Dark-Planet-8B-GGUF] for some writing task(s), and this model for prose specific task(s).
148
 
149
+ <B>Settings: CHAT / ROLEPLAY and/or SMOOTHER operation of this model:</B>
150
+
151
+ In "KoboldCpp" or "oobabooga/text-generation-webui" or "Silly Tavern" ;
152
+
153
+ Set the "Smoothing_factor" to 1.5 to 2.5
154
+
155
+ : in KoboldCpp -> Settings->Samplers->Advanced-> "Smooth_F"
156
+
157
+ : in text-generation-webui -> parameters -> lower right.
158
+
159
+ : In Silly Tavern this is called: "Smoothing"
160
+
161
+
162
+ NOTE: For "text-generation-webui"
163
+
164
+ -> if using GGUFs you need to use "llama_HF" (which involves downloading some config files from the SOURCE version of this model)
165
+
166
+ Source versions (and config files) of my models are here:
167
+
168
+ https://huggingface.co/collections/DavidAU/d-au-source-files-for-gguf-exl2-awq-gptq-hqq-etc-etc-66b55cb8ba25f914cbf210be
169
+
170
+ OTHER OPTIONS:
171
+
172
+ - Increase rep pen to 1.1 to 1.15 (you don't need to do this if you use "smoothing_factor")
173
+
174
+ - If the interface/program you are using to run AI MODELS supports "Quadratic Sampling" ("smoothing") just make the adjustment as noted.
175
+
176
  <B>Model Template:</B>
177
 
178
  This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.