--- license: apache-2.0 tags: - merge - mergekit - lazymergekit - maldv/badger-writer-llama-3-8b - vicgalle/Configurable-Hermes-2-Pro-Llama-3-8B - Orenguteng/Llama-3-8B-Lexi-Uncensored - abacusai/Llama-3-Smaug-8B language: - en base_model: - vicgalle/Configurable-Hermes-2-Pro-Llama-3-8B - maldv/badger-writer-llama-3-8b - Orenguteng/Llama-3-8B-Lexi-Uncensored - abacusai/Llama-3-Smaug-8B library_name: transformers --- # ZeroXClem/Llama-3-Aetheric-Hermes-Lexi-Smaug-8B **Model Architecture:** L3-8B **Model Family:** Llama-3 **Merge Method:** TIES **License:** [Meta Llama Community License](https://llama.meta.com/llama3/license/) --- ### Model Description **Llama-3-Aetheric-Hermes-Lexi-Smaug-8B** is a masterfully merged model designed to offer a balanced, flexible, and high-performing experience. Built on the robust foundation of Llama-3 and drawing from diverse sources, this model is ideal for tasks that require advanced natural language understanding, instruction-following, creative writing, and more. It leverages the unique strengths of each contributing model, yielding a multifaceted, versatile AI capable of adapting to various applications with grace and power. ### Merged Models This model is an amalgamation of the following distinguished lineages: 1. **[maldv/badger-writer-llama-3-8b](https://huggingface.co./maldv/badger-writer-llama-3-8b)**: Known for its expressive storytelling and sci-fi writing capabilities, with a structured approach to immersive text generation. 2. **[vicgalle/Configurable-Hermes-2-Pro-Llama-3-8B](https://huggingface.co./vicgalle/Configurable-Hermes-2-Pro-Llama-3-8B)**: A highly configurable model that enables controlled, safe output with flexible tuning for varying degrees of safety and censorship, perfect for responsible deployments. 3. **[Orenguteng/Llama-3-8B-Lexi-Uncensored](https://huggingface.co./Orenguteng/Llama-3-8B-Lexi-Uncensored)**: A dynamic and unrestricted model designed for candid responses and uncensored creative outputs, empowering more authentic dialogues. 4. **[abacusai/Llama-3-Smaug-8B](https://huggingface.co./abacusai/Llama-3-Smaug-8B)**: Optimized for conversational applications, with enhanced real-world multi-turn dialogue capabilities and stable responsiveness, crucial for interactive and continuous exchanges. ### Merge Configuration This merge was carefully executed using the **TIES** method to combine each model’s unique attributes harmoniously. Here is the YAML configuration used: ```yaml models: - model: maldv/badger-writer-llama-3-8b parameters: density: 0.4 weight: 0.3 # Focused on enhancing storytelling and vivid descriptions - model: vicgalle/Configurable-Hermes-2-Pro-Llama-3-8B parameters: density: 0.6 weight: 0.4 # Prioritizes balanced responses and configurable safety - model: Orenguteng/Llama-3-8B-Lexi-Uncensored parameters: density: 0.5 weight: 0.2 # Adds versatility and creative freedom - model: abacusai/Llama-3-Smaug-8B parameters: density: 0.5 weight: 0.1 # Enhances conversational flow and multi-turn coherence merge_method: ties base_model: Orenguteng/Llama-3-8B-Lexi-Uncensored dtype: bfloat16 parameters: normalize: true out_dtype: float16 ``` ### Key Features - **Storytelling and Creativity:** Leveraging the narrative strength of Badger Writer, this model can generate rich, descriptive stories with intricate character details and immersive settings. - **Configurable Safety and Ethics Compliance:** With Configurable Hermes, users can set desired safety levels, ensuring safe output for sensitive applications or exploring uncensored responses in safe environments. - **Conversational Agility:** Smaug’s conversational tuning makes the model ideal for multi-turn, coherent exchanges, keeping the dialogue engaging and consistent across contexts. - **Unrestricted Dialogue Options:** Lexi provides a flexible, open-ended dialogue experience, perfect for creative brainstorming, research, and unrestricted content generation. ### Ideal Use Cases 1. **Creative Writing and Roleplay:** Perfect for crafting immersive stories, character-driven narratives, and exploratory role-playing scenarios. 2. **Coding and Technical Assistance:** Capable of providing code snippets, explanations, and technical insights, with code generation refined for real-world applications. 3. **Ethical Research and Dialogue:** Ideal for controlled environments where researchers can test response ethics, alignment, and safety controls. 4. **General Instruction-Following:** Useful in assistant-like roles for answering questions, following system prompts, or generating informative content with flexible degrees of guidance. --- ### Quantized Versions For users looking to leverage optimized quantized versions of this model, we are grateful to **MrRaderMacher** for providing GGUF quantizations. These versions are accessible through the links below: - **Static quants**: Available at [huggingface.co/mradermacher/Llama-3-Aetheric-Hermes-Lexi-Smaug-8B-GGUF](https://huggingface.co./mradermacher/Llama-3-Aetheric-Hermes-Lexi-Smaug-8B-GGUF) - **Weighted/imatrix quants**: Available at [huggingface.co/mradermacher/Llama-3-Aetheric-Hermes-Lexi-Smaug-8B-i1-GGUF](https://huggingface.co./mradermacher/Llama-3-Aetheric-Hermes-Lexi-Smaug-8B-i1-GGUF) We recommend users select the **imatrix version** for the best inference performance, as it provides an optimal balance of efficiency and accuracy. Once again, a heartfelt thanks to **MrRaderMacher** for his contributions to the community with these quantized versions. --- ### Example Usage This model has unique Stop Strings, for optimal inference in LM Studio, enter the following setting under Prompt Format > Stop Strings: ```yaml <|reserved_special_token_1|> ``` What are stop strings? Specific strings that when encountered will stop the model from generating more tokens. #### Basic Text Generation ```python from transformers import AutoTokenizer, pipeline import torch model_id = "Llama-3-Aetheric-Hermes-Lexi-Smaug-8B" tokenizer = AutoTokenizer.from_pretrained(model_id) pipe = pipeline( "text-generation", model=model_id, torch_dtype=torch.bfloat16, device_map="auto", ) prompt = "Write a detailed description of a futuristic cityscape at night." outputs = pipe(prompt, max_new_tokens=150, do_sample=True, temperature=0.8) print(outputs[0]['generated_text']) ``` #### Configurable System Prompts (Example for Safe Mode) ```python conversation = [ {"role": "system", "content": "You are a helpful assistant that avoids generating harmful content."}, {"role": "user", "content": "List some fun activities to do in Tokyo."}, ] prompt = tokenizer.apply_chat_template(conversation, tokenize=False, add_generation_prompt=True) outputs = pipe(prompt, max_new_tokens=100) print(outputs[0]["generated_text"]) ``` ### Limitations and Considerations - **Uncensored Outputs:** While powerful, this model can generate uncensored and unrestricted responses. Users should implement additional safety checks or monitor content where needed. - **Bias and Fairness:** Despite combining diverse model sources, this model may carry inherited biases from its training datasets. Users should validate outputs when accuracy and neutrality are essential. - **Computational Requirements:** Optimized for bfloat16, this model may require substantial computational resources. Ensure your setup is equipped for handling the model at its optimal performance. ### Acknowledgments Gratitude goes out to the developers and contributors of the base models: - **Maldv**, for their work on Badger Writer, inspiring vivid storytelling capabilities. - **Vic Gallego**, for their innovation in configurable safety with Hermes, empowering flexible safety settings. - **Orenguteng**, for enabling uncensored conversational capabilities with Lexi. - **Abacus.AI**, for enhancing real-world conversational performance with Smaug. ### License This model adheres to the [Meta Llama Community License](https://llama.meta.com/llama3/license/). Please review the license for detailed usage rights and limitations. --- Enjoy exploring the versatile power of **Llama-3-Aetheric-Hermes-Lexi-Smaug-8B**! Let this model elevate your projects, spark creativity, and unlock new possibilities with every prompt. For additional support, please visit the respective model repositories and join the community conversations to share insights and ask questions. ---