MN-12B-Nymphaea-RP

A fine-tune of Mistral Nemo Instruct 2407 for roleplay and creative writing.

I've trained this mostly for merging with Mistral Nemo fine-tunes.

The SillyTavern preset is available here. For custom presets, please use the Mistral V3-Tekken instruct template.

Training Notes

Trained on the latest iteration of my Darkmere dataset. This version features expanded genre variety, built upon a mix of manually curated synthetics and human-written stories.

The base weights are abliterated via Heretic prior to fine-tuning, so this fine-tune is quite uncensored.

Training Specs

Method:

  • Training Method: DoRA (Weight-Decomposed LoRA)
  • Target Modules all-linear
  • LoRA Rank: 64
  • LoRA Alpha: 64
  • LoRA Dropout: 0.05

Hyperparameters:

  • Batch Size: 2 (Per-device)
  • Gradient Accumulation: 2
  • Epochs: 2
  • Learning Rate: 1e-4
  • Optimizer: adamw_torch_fused
  • LR Scheduler: cosine
  • Noise Level: neftune_noise_alpha=5

Special Thanks

This fine-tune wouldn't be possible without the incredible work of the community:

Downloads last month
54
Safetensors
Model size
12B params
Tensor type
BF16
·
Inference Providers NEW
Input a message to start chatting with 0xA50C1A1/MN-12B-Nymphaea-RP.

Model tree for 0xA50C1A1/MN-12B-Nymphaea-RP

Finetuned
(1)
this model
Quantizations
2 models

Collection including 0xA50C1A1/MN-12B-Nymphaea-RP