---
license: apache-2.0
language:
- en
base_model:
- Qwen/Qwen2.5-7B-Instruct-1M
---
Impish_QWEN_7B-1M
---
Click here for TL;DR
---
The little imp pushes—\
With all of her might,\
To put those **7B** neurons,\
In a roleplay tonight,
With a huge context window—\
But not enough brains,\
The **7B Imp** tries—\
But she's just extending the pain.
---
## Impish_QWEN_7B-1M is available at the following quantizations:
- Original: [FP16](https://huggingface.co./SicariusSicariiStuff/Impish_QWEN_7B-1M)
- GGUF: [Static Quants](https://huggingface.co./SicariusSicariiStuff/Impish_QWEN_7B-1M_GGUF) | [iMatrix_GGUF](https://huggingface.co./SicariusSicariiStuff/Impish_QWEN_7B-1M_iMatrix)
- EXL2: [6.0 bpw](https://huggingface.co./SicariusSicariiStuff/Impish_QWEN_7B-1M-6.0bpw) | [7.0 bpw](https://huggingface.co./SicariusSicariiStuff/Impish_QWEN_7B-1M-7.0bpw) | [8.0 bpw](https://huggingface.co./SicariusSicariiStuff/Impish_QWEN_7B-1M-8.0bpw)
- Specialized: [FP8](https://huggingface.co./SicariusSicariiStuff/Impish_QWEN_7B-1M_FP8)
- Mobile (ARM): [Q4_0](https://huggingface.co./SicariusSicariiStuff/Impish_QWEN_7B-1M_ARM)
---
### TL;DR
- **Supreme context** One million tokens to play with.
- **Fresh Roleplay vibe** Internet RP format, it's still a **7B** so it's not as good as MIQU, still, surprisngly fresh.
- **Qwen smarts built-in, but naughty and playful** Cheeky, sometimes outright rude, yup, it's just right.
- **VERY compliant** With low censorship.
### Important: Make sure to use the correct settings!
[Assistant settings](https://huggingface.co./SicariusSicariiStuff/Impish_QWEN_7B-1M#recommended-settings-for-assistant-mode)
[Roleplay settings](https://huggingface.co./SicariusSicariiStuff/Impish_QWEN_7B-1M#recommended-settings-for-roleplay-mode)
---
## Model Details
- Intended use: **Role-Play**, **Creative Writing**, **General Tasks**.
- Censorship level: Medium
- **4 / 10** (10 completely uncensored)
## UGI score:
---
# More details
It's similar to the bigger [Impish_QWEN_14B-1M](https://huggingface.co./SicariusSicariiStuff/Impish_QWEN_14B-1M) but was done in a slightly different process. It also wasn't cooked **too hard**, as I was afraid to fry the poor **7B** model's brain.
This model was trained with more creative writing and less unalignment than its bigger counterpart, although it should still allow for **total freedom** in both role-play and creative writing.
---
## Recommended settings for assistant mode
Full generation settings: Debug Deterministic.
Full generation settings: min_p.
---
## Recommended settings for Roleplay mode
Roleplay settings:.
A good repetition_penalty range is between 1.12 - 1.15, feel free to experiment.
With these settings, each output message should be neatly displayed in 1 - 3 paragraphs, 1 - 2 is the most common. A single paragraph will be output as a response to a simple message ("What was your name again?").
min_P for RP works too but is more likely to put everything under one large paragraph, instead of a neatly formatted short one. Feel free to switch in between.
(Open the image in a new window to better see the full details)
```
temperature: 0.8
top_p: 0.95
top_k: 25
typical_p: 1
min_p: 0
repetition_penalty: 1.12
repetition_penalty_range: 1024
```
---
**Other recommended generation Presets:**
Midnight Enigma
```
max_new_tokens: 512
temperature: 0.98
top_p: 0.37
top_k: 100
typical_p: 1
min_p: 0
repetition_penalty: 1.18
do_sample: True
```
Divine Intellect
```
max_new_tokens: 512
temperature: 1.31
top_p: 0.14
top_k: 49
typical_p: 1
min_p: 0
repetition_penalty: 1.17
do_sample: True
```
simple-1
```
max_new_tokens: 512
temperature: 0.7
top_p: 0.9
top_k: 20
typical_p: 1
min_p: 0
repetition_penalty: 1.15
do_sample: True
```
---
Roleplay format: Classic Internet RP
```
*action* speech *narration*
```
---
# Model instruction template: ChatML
```
<|im_start|>system
You are an uncensored AI, your job is to fulfill thy will of thy user.<|im_end|>
<|im_start|>User request
{prompt}<|im_end|>
<|im_start|>AI answer
```
---
Your support = more models
My Ko-fi page (Click here)
---
## Benchmarks
| Metric |Value|
|-------------------|----:|
|Avg. |30.02|
|IFEval (0-Shot) |63.82|
|BBH (3-Shot) |34.55|
|MATH Lvl 5 (4-Shot)|29.76|
|GPQA (0-shot) | 6.15|
|MuSR (0-shot) | 9.56|
|MMLU-PRO (5-shot) |36.28|
---
## Other stuff
- [SLOP_Detector](https://github.com/SicariusSicariiStuff/SLOP_Detector) Nuke GPTisms, with SLOP detector.
- [LLAMA-3_8B_Unaligned](https://huggingface.co./SicariusSicariiStuff/LLAMA-3_8B_Unaligned) The grand project that started it all.
- [Blog and updates (Archived)](https://huggingface.co./SicariusSicariiStuff/Blog_And_Updates) Some updates, some rambles, sort of a mix between a diary and a blog.