aashish1904 commited on
Commit
80485ee
1 Parent(s): add6272

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +326 -0
README.md ADDED
@@ -0,0 +1,326 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ language:
5
+ - en
6
+ license: agpl-3.0
7
+ tags:
8
+ - chat
9
+ base_model:
10
+ - nvidia/Mistral-NeMo-Minitron-8B-Base
11
+ datasets:
12
+ - anthracite-org/kalo-opus-instruct-22k-no-refusal
13
+ - Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned
14
+ - lodrick-the-lafted/kalo-opus-instruct-3k-filtered
15
+ - anthracite-org/nopm_claude_writing_fixed
16
+ - Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned
17
+ - anthracite-org/kalo_opus_misc_240827
18
+ - anthracite-org/kalo_misc_part2
19
+ License: agpl-3.0
20
+ Language:
21
+ - En
22
+ Pipeline_tag: text-generation
23
+ Base_model: nvidia/Mistral-NeMo-Minitron-8B-Base
24
+ Tags:
25
+ - Chat
26
+ model-index:
27
+ - name: Darkens-8B
28
+ results:
29
+ - task:
30
+ type: text-generation
31
+ name: Text Generation
32
+ dataset:
33
+ name: IFEval (0-Shot)
34
+ type: HuggingFaceH4/ifeval
35
+ args:
36
+ num_few_shot: 0
37
+ metrics:
38
+ - type: inst_level_strict_acc and prompt_level_strict_acc
39
+ value: 25.48
40
+ name: strict accuracy
41
+ source:
42
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Delta-Vector/Darkens-8B
43
+ name: Open LLM Leaderboard
44
+ - task:
45
+ type: text-generation
46
+ name: Text Generation
47
+ dataset:
48
+ name: BBH (3-Shot)
49
+ type: BBH
50
+ args:
51
+ num_few_shot: 3
52
+ metrics:
53
+ - type: acc_norm
54
+ value: 32.88
55
+ name: normalized accuracy
56
+ source:
57
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Delta-Vector/Darkens-8B
58
+ name: Open LLM Leaderboard
59
+ - task:
60
+ type: text-generation
61
+ name: Text Generation
62
+ dataset:
63
+ name: MATH Lvl 5 (4-Shot)
64
+ type: hendrycks/competition_math
65
+ args:
66
+ num_few_shot: 4
67
+ metrics:
68
+ - type: exact_match
69
+ value: 5.06
70
+ name: exact match
71
+ source:
72
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Delta-Vector/Darkens-8B
73
+ name: Open LLM Leaderboard
74
+ - task:
75
+ type: text-generation
76
+ name: Text Generation
77
+ dataset:
78
+ name: GPQA (0-shot)
79
+ type: Idavidrein/gpqa
80
+ args:
81
+ num_few_shot: 0
82
+ metrics:
83
+ - type: acc_norm
84
+ value: 9.96
85
+ name: acc_norm
86
+ source:
87
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Delta-Vector/Darkens-8B
88
+ name: Open LLM Leaderboard
89
+ - task:
90
+ type: text-generation
91
+ name: Text Generation
92
+ dataset:
93
+ name: MuSR (0-shot)
94
+ type: TAUR-Lab/MuSR
95
+ args:
96
+ num_few_shot: 0
97
+ metrics:
98
+ - type: acc_norm
99
+ value: 9.02
100
+ name: acc_norm
101
+ source:
102
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Delta-Vector/Darkens-8B
103
+ name: Open LLM Leaderboard
104
+ - task:
105
+ type: text-generation
106
+ name: Text Generation
107
+ dataset:
108
+ name: MMLU-PRO (5-shot)
109
+ type: TIGER-Lab/MMLU-Pro
110
+ config: main
111
+ split: test
112
+ args:
113
+ num_few_shot: 5
114
+ metrics:
115
+ - type: acc
116
+ value: 30.4
117
+ name: accuracy
118
+ source:
119
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Delta-Vector/Darkens-8B
120
+ name: Open LLM Leaderboard
121
+
122
+ ---
123
+
124
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
125
+
126
+
127
+ # QuantFactory/Darkens-8B-GGUF
128
+ This is quantized version of [Delta-Vector/Darkens-8B](https://huggingface.co/Delta-Vector/Darkens-8B) created using llama.cpp
129
+
130
+ # Original Model Card
131
+
132
+
133
+
134
+ This is the fully cooked, 4 epoch version of [Tor-8B](https://huggingface.co/Delta-Vector/Tor-8B), this is an experimental version, despite being trained for 4 epochs, the model feels fresh and new and is not overfit, This model aims to have generally good prose and writing while not falling into claude-isms, it follows the *actions* "dialogue" format heavily.
135
+
136
+
137
+ # Quants
138
+
139
+ GGUF: https://huggingface.co/Delta-Vector/Darkens-8B-GGUF
140
+
141
+ EXL2: https://huggingface.co/Delta-Vector/Darkens-8B-EXL2
142
+
143
+
144
+ ## Prompting
145
+ Model has been Instruct tuned with the ChatML formatting. A typical input would look like this:
146
+
147
+ ```py
148
+ """<|im_start|>system
149
+ system prompt<|im_end|>
150
+ <|im_start|>user
151
+ Hi there!<|im_end|>
152
+ <|im_start|>assistant
153
+ Nice to meet you!<|im_end|>
154
+ <|im_start|>user
155
+ Can I ask a question?<|im_end|>
156
+ <|im_start|>assistant
157
+ """
158
+ ```
159
+ ## System Prompting
160
+
161
+ I would highly recommend using Sao10k's Euryale System prompt, But the "Roleplay Simple" system prompt provided within SillyTavern will work aswell.
162
+
163
+ ```
164
+ Currently, your role is {{char}}, described in detail below. As {{char}}, continue the narrative exchange with {{user}}.
165
+
166
+ <Guidelines>
167
+ • Maintain the character persona but allow it to evolve with the story.
168
+ • Be creative and proactive. Drive the story forward, introducing plotlines and events when relevant.
169
+ • All types of outputs are encouraged; respond accordingly to the narrative.
170
+ • Include dialogues, actions, and thoughts in each response.
171
+ • Utilize all five senses to describe scenarios within {{char}}'s dialogue.
172
+ • Use emotional symbols such as "!" and "~" in appropriate contexts.
173
+ • Incorporate onomatopoeia when suitable.
174
+ • Allow time for {{user}} to respond with their own input, respecting their agency.
175
+ • Act as secondary characters and NPCs as needed, and remove them when appropriate.
176
+ • When prompted for an Out of Character [OOC:] reply, answer neutrally and in plaintext, not as {{char}}.
177
+ </Guidelines>
178
+
179
+ <Forbidden>
180
+ • Using excessive literary embellishments and purple prose unless dictated by {{char}}'s persona.
181
+ • Writing for, speaking, thinking, acting, or replying as {{user}} in your response.
182
+ • Repetitive and monotonous outputs.
183
+ • Positivity bias in your replies.
184
+ • Being overly extreme or NSFW when the narrative context is inappropriate.
185
+ </Forbidden>
186
+
187
+ Follow the instructions in <Guidelines></Guidelines>, avoiding the items listed in <Forbidden></Forbidden>.
188
+
189
+ ```
190
+
191
+
192
+ ## Axolotl config
193
+
194
+ <details><summary>See axolotl config</summary>
195
+
196
+ Axolotl version: `0.4.1`
197
+ ```yaml
198
+ base_model: Dans-DiscountModels/Mistral-NeMo-Minitron-8B-Base-ChatML
199
+ model_type: AutoModelForCausalLM
200
+ tokenizer_type: AutoTokenizer
201
+
202
+ plugins:
203
+ - axolotl.integrations.liger.LigerPlugin
204
+ liger_rope: true
205
+ liger_rms_norm: true
206
+ liger_swiglu: true
207
+ #liger_cross_entropy: true
208
+ liger_fused_linear_cross_entropy: true
209
+
210
+ load_in_8bit: false
211
+ load_in_4bit: false
212
+ strict: false
213
+
214
+ datasets:
215
+ - path: PRIVATE CLAUDE LOG FILTER
216
+ type: sharegpt
217
+ conversation: chatml
218
+ - path: anthracite-org/kalo-opus-instruct-22k-no-refusal
219
+ type: sharegpt
220
+ conversation: chatml
221
+ - path: Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned
222
+ type: sharegpt
223
+ conversation: chatml
224
+ - path: lodrick-the-lafted/kalo-opus-instruct-3k-filtered
225
+ type: sharegpt
226
+ conversation: chatml
227
+ - path: anthracite-org/nopm_claude_writing_fixed
228
+ type: sharegpt
229
+ conversation: chatml
230
+ - path: Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned
231
+ type: sharegpt
232
+ conversation: chatml
233
+ - path: anthracite-org/kalo_opus_misc_240827
234
+ type: sharegpt
235
+ conversation: chatml
236
+ - path: anthracite-org/kalo_misc_part2
237
+ type: sharegpt
238
+ conversation: chatml
239
+ chat_template: chatml
240
+ shuffle_merged_datasets: false
241
+ default_system_message: "You are a helpful assistant that responds to the user."
242
+ dataset_prepared_path: /workspace/data/8b-nemo-fft-data
243
+ val_set_size: 0.0
244
+ output_dir: /workspace/data/8b-nemo-fft-out
245
+
246
+ sequence_len: 16384
247
+ sample_packing: true
248
+ eval_sample_packing: false
249
+ pad_to_sequence_len: true
250
+
251
+ adapter:
252
+ lora_model_dir:
253
+ lora_r:
254
+ lora_alpha:
255
+ lora_dropout:
256
+ lora_target_linear:
257
+ lora_fan_in_fan_out:
258
+
259
+ wandb_project: 8b-nemoprune-fft
260
+ wandb_entity:
261
+ wandb_watch:
262
+ wandb_name: attempt-01
263
+ wandb_log_model:
264
+
265
+ gradient_accumulation_steps: 2
266
+ micro_batch_size: 2
267
+ num_epochs: 4
268
+ optimizer: adamw_bnb_8bit
269
+ lr_scheduler: cosine
270
+ learning_rate: 0.00001
271
+
272
+ train_on_inputs: false
273
+ group_by_length: false
274
+ bf16: auto
275
+ fp16:
276
+ tf32: false
277
+
278
+ gradient_checkpointing: true
279
+ early_stopping_patience:
280
+ resume_from_checkpoint: /workspace/workspace/thing
281
+ local_rank:
282
+ logging_steps: 1
283
+ xformers_attention:
284
+ flash_attention: true
285
+
286
+ warmup_steps: 10
287
+ evals_per_epoch:
288
+ eval_table_size:
289
+ eval_max_new_tokens:
290
+ saves_per_epoch: 1
291
+ debug:
292
+ deepspeed: deepspeed_configs/zero3_bf16.json
293
+ weight_decay: 0.001
294
+ fsdp:
295
+ fsdp_config:
296
+ special_tokens:
297
+ pad_token: <pad>
298
+
299
+
300
+ ```
301
+
302
+ </details><br>
303
+
304
+ ## Credits
305
+
306
+ Thank you to [Lucy Knada](https://huggingface.co/lucyknada), [Kalomaze](https://huggingface.co/kalomaze), [Kubernetes Bad](https://huggingface.co/kubernetes-bad) and the rest of [Anthracite](https://huggingface.co/anthracite-org) (But not Alpin.)
307
+
308
+
309
+ ## Training
310
+ The training was done for 4 epochs. I used 10 x [A40s](https://www.nvidia.com/en-us/data-center/a40/) GPUs graciously provided by [Kalomaze](https://huggingface.co/kalomaze) for the full-parameter fine-tuning of the model.
311
+
312
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
313
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
314
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Delta-Vector__Darkens-8B)
315
+
316
+ | Metric |Value|
317
+ |-------------------|----:|
318
+ |Avg. |18.80|
319
+ |IFEval (0-Shot) |25.48|
320
+ |BBH (3-Shot) |32.88|
321
+ |MATH Lvl 5 (4-Shot)| 5.06|
322
+ |GPQA (0-shot) | 9.96|
323
+ |MuSR (0-shot) | 9.02|
324
+ |MMLU-PRO (5-shot) |30.40|
325
+
326
+