Static Quants:
https://huggingface.co/IggyLux/MN-VelvetCafe-RP-12B-Q4_K_M-GGUF https://huggingface.co/IggyLux/MN-VelvetCafe-RP-12B-Q8_0-GGUF
This is my 4th Attempt at a merge of finetunes and the only one I've been happy with. I'm always looking for new merges/finetunes of 12b's due to my 8gb VRAM limitations so I decided to merge my own. I focus mainly on Group Chat RP's personally so when I RP it's mostly +2 Characters if not more.
My take at what I think makes this merged finetune model good:
- π Strong scene/position/clothing tracking for immersive multi-turn RP
- β€οΈ Balanced emotional responses β no sudden aggression or refusal spikes unless fitting the narrative of RP (sometimes due to relations you might want this type of response)
- π Handles author's notes/system prompts reliably
My Goal was to take Dans PE hoping that it's character/clothes/personality tracking and consistency would shine when combined with Neona. Neona is really good at adapting to writing styles and instruction following from my experience using it as a daily driver. Combining the two resulted in very good visual focused RP.
I dislike when models forget clothing, positioning and don't reply in responses detailing changes like that. This often leads to models hallucinating/forgetting positions and clothing specifics that breaks immersion for me. This model seems to feel more visually detailed and descriptive and aware of some of the better things Dan does while keeping some of the instruction following and closer to neurtral emotional responses of Neona.
I encourage you to try both down below as I really love these models. Thank you for making them @kyx0r and @PocketDoc
Dans-PersonalityEngine-V1.3.0-12b is one of those local models that just clicks really well for roleplay. The creators tuned it hard on a ton of different datasets, and they made sure roleplay and creative writing were right up there as core strengths, not some side feature tacked on. That means it naturally picks up on writing good dialogue, keeping descriptions flowing, and building scenes that feel alive instead of stiff or robotic. Unfortunately it looks like it was created before tokenizer issues with mistral nemo were fixed. Due to that it might have format/puncuation issues that sometimes can be a bit annoying. It also has a tendency to favor shorter replies.
Neona-12B is a personal favorite of mine. It's a model that seems unbiased in roleplay. If you want slice of life and keeping things SFW it really adapts well. If you want NSFW ERP it can adjust and adapt to that as well too. The model doesn't seem to jump you based on subtle contact or act in extremes like some other finetunes do. I feel like it has a stability emotionally most models don't have. It also seems to handle system prompts/authors notes and instructions well which not all models do.
My preferred format for Roleplaying in Sillytavern is:
- ChatML or
- Mistral V3-Tekken
My sampler settings for Text Completion preset are included as well with the model, though I personally believe you should find what you like best yourself instead of relying on others. But if you need it, feel free to use it as a place to start: https://huggingface.co/IggyLux/MN-VelvetCafe-RP-12B/blob/main/Iggy's-RP-Preset.json
It's set to 8192 context. Which is a great starting point for 8gb VRAM users and with 356 response length to conserve context, I tweak it to 512 for more detail and 1024 for scene climaxes in great detail.
My preset temp is 0.8 if for some reason you want it to be less creative or more grounded you can go as low as 0.4 (play around with it)
My setup is:
- KoboldCpp GUI for the backend GGUF model loading found here: https://github.com/LostRuins/koboldcpp
- Sillytavern for the front end chat interface https://github.com/SillyTavern/SillyTavern (current version 1.16.0)
- π https://github.com/aikohanasaki/SillyTavern-MemoryBooks/ - For keeping context low and saving older responses as Memories in a Lore Book
- π https://github.com/leandrojofre/SillyTavern-Presence - For Group Chats: Using Presence lets you select what characters can see the user and char's messages.
- π£οΈ https://github.com/mattjaybe/SillyTavern-EchoChamber - New* I recently found this and thought it was pretty cool, you can have a chat comment on your RP.
- π― https://github.com/Samueras/GuidedGenerations-Extension - Helps steer stubborn models, use guides to lock in scenes/details/clothing/positions and more.
Character Cards and Roleplay Usage/Examples: For some reason a lot of people do things differently (usually based on old tutorials) but I refrain from using opening messages on character cards, example dialogue and things that would sway the model to speak for the user. I also make my own characters after using Chub/Venus/Playground seeing how there's a lot of 1500-2000 token character cards with p-lists like this:
If you see a character card like this it might work, but honestly it's not really neccesary to format using P lists and other stuff like that now days. Models can read standard text formatting just fine.
Another thing I try to avoid doing or downloading is character cards that use example dialogue, especially one's that speak for the user in examples:
As you can see in this example the creator has mostly example dialogue between a "Interviewer" and the Jinn. This kind of example formatting might lead to the introduction of a 3rd character and start speaking for the "interviewer" or maybe even talk for your character in that way as well.
I was helping someone troubleshoot using the model and decided (after just waking up without my first cup of coffee so ignore some mistakes in text!) and figured I'd share how I structure my roleplays in sillytavern:
I'm officially naming this the Iggy format since I don't see anyone else start RP's in sillytavern this way.
For example I'm not a big fan of Character Cards starting the scenario with a first message (tends to set bad habits) so I'll either let my character open up with the first Message and a lot of times before that first message I'll send a system level prompt using /sys to set the scenario in some format of narrative summary.
If for some reason this leaves context due to limits I'll repurpose my opening scenario into the group chat scenario here:
Or pause the summary extenstion and paste it into there, another option is tossing part of it into authors notes (condensed version to save context) or if you are using memory books extension suggested above, it will generally be included in it's worldinfo lore books entry it creates.
I do my best to share what I've learned having technical limitations over the last few years roleplaying, if you have any issues or problems feel free to ask and I'll try to help!
This is a merge of pre-trained language models created using mergekit.
Merge Details
Merge Method
This model was merged using the SLERP merge method.
Models Merged
The following models were included in the merge:
Configuration
The following YAML configuration was used to produce this model:
models:
- model: kyx0r/Neona-12B
- model: PocketDoc/Dans-PersonalityEngine-V1.3.0-12b
merge_method: slerp
base_model: kyx0r/Neona-12B
parameters:
t:
- value: 0.2
- filter: self_attn
value: [0, 0.2, 0.4, 0.6, 0.8, 1]
- filter: mlp
value: [1, 0.8, 0.6, 0.4, 0.2, 0]
dtype: bfloat16
chat_template: "chatml"
tokenizer:
source: "base"
- Downloads last month
- 567




