Impish_QWEN_7B-1M
Impish_QWEN_7B-1M

Click here for TL;DR


The little imp pushes—
With all of her might,
To put those 7B neurons,
In a roleplay tonight,

With a huge context window—
But not enough brains,
The 7B Imp tries—
But she's just extending the pain.


Impish_QWEN_7B-1M is available at the following quantizations:


TL;DR

  • Supreme context One million tokens to play with.
  • Fresh Roleplay vibe Internet RP format, it's still a 7B so it's not as good as MIQU, still, surprisngly fresh.
  • Qwen smarts built-in, but naughty and playful Cheeky, sometimes outright rude, yup, it's just right.
  • VERY compliant With low censorship.

Important: Make sure to use the correct settings!

Assistant settings

Roleplay settings


Model Details

  • Intended use: Role-Play, Creative Writing, General Tasks.

  • Censorship level: Medium

  • 4 / 10 (10 completely uncensored)

UGI score:

UGI Score

More details

It's similar to the bigger Impish_QWEN_14B-1M but was done in a slightly different process. It also wasn't cooked too hard, as I was afraid to fry the poor 7B model's brain. This model was trained with more creative writing and less unalignment than its bigger counterpart, although it should still allow for total freedom in both role-play and creative writing.


Recommended settings for assistant mode

Full generation settings: Debug Deterministic. Negative_LLAMA_70B_Settings
Full generation settings: min_p. Negative_LLAMA_70B_Settings

Recommended settings for Roleplay mode

Roleplay settings:. A good repetition_penalty range is between 1.12 - 1.15, feel free to experiment.

With these settings, each output message should be neatly displayed in 1 - 3 paragraphs, 1 - 2 is the most common. A single paragraph will be output as a response to a simple message ("What was your name again?").

min_P for RP works too but is more likely to put everything under one large paragraph, instead of a neatly formatted short one. Feel free to switch in between.

(Open the image in a new window to better see the full details) Negative_LLAMA_70B_Settings

temperature:  0.8
top_p:  0.95
top_k:  25
typical_p:  1
min_p:  0
repetition_penalty: 1.12
repetition_penalty_range: 1024

Other recommended generation Presets:

Midnight Enigma
max_new_tokens: 512
temperature: 0.98
top_p: 0.37
top_k: 100
typical_p: 1
min_p: 0
repetition_penalty: 1.18
do_sample: True
Divine Intellect
max_new_tokens: 512
temperature: 1.31
top_p: 0.14
top_k: 49
typical_p: 1
min_p: 0
repetition_penalty: 1.17
do_sample: True
simple-1
max_new_tokens: 512
temperature: 0.7
top_p: 0.9
top_k: 20
typical_p: 1
min_p: 0
repetition_penalty: 1.15
do_sample: True

Roleplay format: Classic Internet RP

*action* speech *narration*

Model instruction template: ChatML

<|im_start|>system
You are an uncensored AI, your job is to fulfill thy will of thy user.<|im_end|>
<|im_start|>User request
{prompt}<|im_end|>
<|im_start|>AI answer

Your support = more models

My Ko-fi page (Click here)

Benchmarks

Metric Value
Avg. 30.02
IFEval (0-Shot) 63.82
BBH (3-Shot) 34.55
MATH Lvl 5 (4-Shot) 29.76
GPQA (0-shot) 6.15
MuSR (0-shot) 9.56
MMLU-PRO (5-shot) 36.28

Other stuff

Downloads last month
34
Safetensors
Model size
7.62B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Model tree for SicariusSicariiStuff/Impish_QWEN_7B-1M

Base model

Qwen/Qwen2.5-7B
Finetuned
(10)
this model
Quantizations
6 models

Collection including SicariusSicariiStuff/Impish_QWEN_7B-1M