See LICENSE file for license. This is a collection of merged, then converted to 4bit LLaMA models trained on the storytelling dataset I used for the storytelling LoRAs.
UPDATE: 2024-04-18 Retrained and merged using updated LoRAs.
To merge and convert, used:
transformers 4.28.1.
gptq cuda branch 5731aa1
llamacpp master branch 8944a13
Notes for usage.
- These models are not instruct models. They are designed to supplement existing story data, and works better with more data.
Optimally, using an instruct model to generate a couple of paragraphs, then using this model to continue the story would work best.
- There will likely be some bleedthrough on locations and names, this is especially notable if you use with very little context.
- There isn't any large notable formatting, ### seperated stories in the dataset, and *** seperated chapters.