LuxBannerVelvetCafe

Velvet_Cafe

Static Quants:

https://huggingface.co/IggyLux/MN-VelvetCafe-RP-12B-Q4_K_M-GGUF https://huggingface.co/IggyLux/MN-VelvetCafe-RP-12B-Q8_0-GGUF

This is my 4th Attempt at a merge of finetunes and the only one I've been happy with. I'm always looking for new merges/finetunes of 12b's due to my 8gb VRAM limitations so I decided to merge my own. I focus mainly on Group Chat RP's personally so when I RP it's mostly +2 Characters if not more.

My take at what I think makes this merged finetune model good:

  • 🌟 Strong scene/position/clothing tracking for immersive multi-turn RP
  • ❀️ Balanced emotional responses β€” no sudden aggression or refusal spikes unless fitting the narrative of RP (sometimes due to relations you might want this type of response)
  • πŸ“ Handles author's notes/system prompts reliably

My Goal was to take Dans PE hoping that it's character/clothes/personality tracking and consistency would shine when combined with Neona. Neona is really good at adapting to writing styles and instruction following from my experience using it as a daily driver. Combining the two resulted in very good visual focused RP.

I dislike when models forget clothing, positioning and don't reply in responses detailing changes like that. This often leads to models hallucinating/forgetting positions and clothing specifics that breaks immersion for me. This model seems to feel more visually detailed and descriptive and aware of some of the better things Dan does while keeping some of the instruction following and closer to neurtral emotional responses of Neona.

I encourage you to try both down below as I really love these models. Thank you for making them @kyx0r and @PocketDoc

  • Dans-PersonalityEngine-V1.3.0-12b is one of those local models that just clicks really well for roleplay. The creators tuned it hard on a ton of different datasets, and they made sure roleplay and creative writing were right up there as core strengths, not some side feature tacked on. That means it naturally picks up on writing good dialogue, keeping descriptions flowing, and building scenes that feel alive instead of stiff or robotic. Unfortunately it looks like it was created before tokenizer issues with mistral nemo were fixed. Due to that it might have format/puncuation issues that sometimes can be a bit annoying. It also has a tendency to favor shorter replies.

  • Neona-12B is a personal favorite of mine. It's a model that seems unbiased in roleplay. If you want slice of life and keeping things SFW it really adapts well. If you want NSFW ERP it can adjust and adapt to that as well too. The model doesn't seem to jump you based on subtle contact or act in extremes like some other finetunes do. I feel like it has a stability emotionally most models don't have. It also seems to handle system prompts/authors notes and instructions well which not all models do.

My preferred format for Roleplaying in Sillytavern is:

  • ChatML or
  • Mistral V3-Tekken

SamplerSettingsModelButton200

My sampler settings for Text Completion preset are included as well with the model, though I personally believe you should find what you like best yourself instead of relying on others. But if you need it, feel free to use it as a place to start: https://huggingface.co/IggyLux/MN-VelvetCafe-RP-12B/blob/main/Iggy's-RP-Preset.json

It's set to 8192 context. Which is a great starting point for 8gb VRAM users and with 356 response length to conserve context, I tweak it to 512 for more detail and 1024 for scene climaxes in great detail.

My preset temp is 0.8 if for some reason you want it to be less creative or more grounded you can go as low as 0.4 (play around with it)

My setup is:

ModelButtonIL200


Character Cards and Roleplay Usage/Examples: For some reason a lot of people do things differently (usually based on old tutorials) but I refrain from using opening messages on character cards, example dialogue and things that would sway the model to speak for the user. I also make my own characters after using Chub/Venus/Playground seeing how there's a lot of 1500-2000 token character cards with p-lists like this:

image If you see a character card like this it might work, but honestly it's not really neccesary to format using P lists and other stuff like that now days. Models can read standard text formatting just fine.

Another thing I try to avoid doing or downloading is character cards that use example dialogue, especially one's that speak for the user in examples:

image As you can see in this example the creator has mostly example dialogue between a "Interviewer" and the Jinn. This kind of example formatting might lead to the introduction of a 3rd character and start speaking for the "interviewer" or maybe even talk for your character in that way as well.

I was helping someone troubleshoot using the model and decided (after just waking up without my first cup of coffee so ignore some mistakes in text!) and figured I'd share how I structure my roleplays in sillytavern:

image I'm officially naming this the Iggy format since I don't see anyone else start RP's in sillytavern this way.

For example I'm not a big fan of Character Cards starting the scenario with a first message (tends to set bad habits) so I'll either let my character open up with the first Message and a lot of times before that first message I'll send a system level prompt using /sys to set the scenario in some format of narrative summary.

If for some reason this leaves context due to limits I'll repurpose my opening scenario into the group chat scenario here:

image

Or pause the summary extenstion and paste it into there, another option is tossing part of it into authors notes (condensed version to save context) or if you are using memory books extension suggested above, it will generally be included in it's worldinfo lore books entry it creates.

I do my best to share what I've learned having technical limitations over the last few years roleplaying, if you have any issues or problems feel free to ask and I'll try to help!

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the SLERP merge method.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: kyx0r/Neona-12B
  - model: PocketDoc/Dans-PersonalityEngine-V1.3.0-12b
merge_method: slerp
base_model: kyx0r/Neona-12B
parameters:
  t:
    - value: 0.2
    - filter: self_attn
      value: [0, 0.2, 0.4, 0.6, 0.8, 1]
    - filter: mlp
      value: [1, 0.8, 0.6, 0.4, 0.2, 0]
dtype: bfloat16
chat_template: "chatml"
tokenizer:
  source: "base"
Downloads last month
567
Safetensors
Model size
12B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for IggyLux/MN-VelvetCafe-RP-12B