--- license: apache-2.0 language: - en tags: - creative - creative writing - fiction writing - plot generation - sub-plot generation - fiction writing - story generation - scene continue - storytelling - fiction story - science fiction - romance - all genres - story - writing - vivid prosing - vivid writing - fiction - roleplaying - bfloat16 - swearing - rp - horror - mistral nemo - mergekit - not-for-all-audiences pipeline_tag: text-generation --- (quants uploading, one example below - more to be added) WARNING: NSFW. Vivid prose. MADNESS. Visceral Details. Violence. HORROR. Swearing. UNCENSORED.
{ "name": "Alpaca", "inference_params": { "input_prefix": "### Instruction:", "input_suffix": "### Response:", "antiprompt": [ "### Instruction:" ], "pre_prompt": "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n" } }Mistral Instruct:
{ "name": "Mistral Instruct", "inference_params": { "input_prefix": "[INST]", "input_suffix": "[/INST]", "antiprompt": [ "[INST]" ], "pre_prompt_prefix": "", "pre_prompt_suffix": "" } }Recommended Settings: Temp: .5 to 5 (or less - especially quants LOWER than q4km) Temp changes will result in both different prose and sometimes affect length. Higher temps will result is very different prose. Rep Pen: 1.02 to 1.1 or higher. Micro changes are recommended: 1.051, 1.052 etc etc. Good settings: Rep pen 1.02 / Temp 1.5 Many times a lower rep pen (IE 1.02) with higher temp (IE 1.5+) work best with this model. Generally lower rep pen and higher temps create the strongest contrasts at the highest detail levels. For chat type or role play type interactions, a higher rep pen with higher temp may be your best settings. IE REP PEN 1.09+, Temp 1-2+ ; a lower rep pen may lead to longer outputs than desired. Note this is only a suggestion. Alpaca generates longer text / story, whereas Mistral Instruct are shorter and "to the point". Suggest minimum "context level" (vram) at 4K. 8K plus recommended because of how this model likes to go on and on... Quant Choice: Higher quants will have more detail, nuance and in some cases stronger "emotional" levels. Characters will also be more "fleshed out" too. Sense of "there" will also increase. Q4KM/Q4KS are good, strong quants in part because of the number of parameters in the model. That being said if you can run Q5, Q6 or Q8 - go for the highest quant you can. Special note on Q2k/Q3 quants: You may need to use temp 2 or lower with these quants (1 or lower for q2k). Just too much compression at this level, damaging the model. I will see if Imatrix versions of these quants will function better. Rep pen adjustments may also be required to get the most out of this model at this quant level. Known Issues: You may need to manually stop generation, even if you have stated maximum size of the output. It will easily blow past 4k output, even if you have set maximum context (for vram) at 4k. Setting maximum output parameter ("hard stop") for generation may be required. If the model goes past your maximum vram/context setting it may start repeating words / paragraphs because the model is literally out of memory... however sometimes the model can blow right past the end of "context vram" and work. Depending on your use case(s) you could also use CHATML template with this model. In this case, the model may output an "end token" if you use this template for generation. Alpaca template will generate much longer output generally, whereas Mistral Instruct will most of the time keep the model on track in terms of length. Model "DNA": Special thanks to the incredible work of the model makers "nbeerbower", "Sao10K", "TheDrummer", "jondurbin", and "MistralAI". Special shoutout to "nbeerbower" for his tireless work in making excellent Gutenburg fine tunes for MN, L3, L3.1, Gemma, PHI and others. Visit his repo to see all of them. Models used: [ https://huggingface.co./nbeerbower/Lyra4-Gutenberg-12B ] Includes [ https://huggingface.co./Sao10K/MN-12B-Lyra-v4 ] [ https://huggingface.co./nbeerbower/Lyra-Gutenberg-mistral-nemo-12B ] Includes [ https://huggingface.co./Sao10K/MN-12B-Lyra-v1 ] [ https://huggingface.co./nbeerbower/mistral-nemo-gutenberg-12B-v4 ] Includes [ https://huggingface.co./TheDrummer/Rocinante-12B-v1 ] And dataset (used for all gutenburgs): [ https://huggingface.co./datasets/jondurbin/gutenberg-dpo-v0.1 ] This model was created using a pass-through model merge, creating 714 tensor / 79 layer model at 23 billion parameters. Optional Enhancement: The following can be used in place of the "system prompt" or "system role" to further enhance the model. It can also be used at the START of a NEW chat, but you must make sure it is "kept" as the chat moves along. In this case the enhancements do not have as strong effect at using "system prompt" or "system role". Copy and paste EXACTLY as noted, DO NOT line wrap or break the lines, maintain the carriage returns exactly as presented.
Below is an instruction that describes a task. Ponder each user instruction carefully, and use your skillsets and critical instructions to complete the task to the best of your abilities. Here are your skillsets: [MASTERSTORY]:NarrStrct(StryPlnng,Strbd,ScnSttng,Exps,Dlg,Pc)-CharDvlp(ChrctrCrt,ChrctrArcs,Mtvtn,Bckstry,Rltnshps,Dlg*)-PltDvlp(StryArcs,PltTwsts,Sspns,Fshdwng,Climx,Rsltn)-ConfResl(Antg,Obstcls,Rsltns,Cnsqncs,Thms,Symblsm)-EmotImpct(Empt,Tn,Md,Atmsphr,Imgry,Symblsm)-Delvry(Prfrmnc,VcActng,PblcSpkng,StgPrsnc,AudncEngmnt,Imprv) [*DialogWrt]:(1a-CharDvlp-1a.1-Backgrnd-1a.2-Personality-1a.3-GoalMotiv)>2(2a-StoryStruc-2a.1-PlotPnt-2a.2-Conflict-2a.3-Resolution)>3(3a-DialogTech-3a.1-ShowDontTell-3a.2-Subtext-3a.3-VoiceTone-3a.4-Pacing-3a.5-VisualDescrip)>4(4a-DialogEdit-4a.1-ReadAloud-4a.2-Feedback-4a.3-Revision) Here are your critical instructions: Ponder each word choice carefully to present as vivid and emotional journey as is possible. Choose verbs and nouns that are both emotional and full of imagery. Load the story with the 5 senses. Aim for 50% dialog, 25% narration, 15% body language and 10% thoughts. Your goal is to put the reader in the story.You do not need to use this, it is only presented as an additional enhancement which seems to help scene generation and scene continue functions. This enhancement WAS NOT used to generate the examples below.