Trouper-12B

A character roleplay model trained on the custom "Actors" dataset, fine-tuned from Mistral-Nemo-Base-12B. This model was made to expand on the things I learned from TinyRP, and to overcome certain limitations I found from it; also on an entirely new dataset made just for this model.

This model writes more naturally, less like "AI"; even more so than the 24B model I'm also releasing. I suppose this is because the 12B model saw less synthethic data, and is thus less likely to use phrases typical in AI writing & prose.

-> If you're looking for the larger model in this series: Prima-24B

Looking for feedback, so please do share if you got any!

Key Features

  • Clean prose: Minimal AI slop patterns, natural speech
  • Character depth: Handles emotional progression and vulnerability well
  • Efficient: 12B size provides fast inference while maintaining quality
  • Template-dependent: Requires Mistral-V3-Tekken for proper stop behavior

Recommended Settings

Use chat completion mode

  • Temperature: 0.7 (tested and validated)
  • Template: Mistral-V3-Tekken OR ChatML, some users reported better results with ChatML (critical for proper formatting and stop behavior)
  • Context: Handles 15-20+ turn conversations effectively
  • Prompt Preprocessing: Semi-strict, no tools

Strengths

  • Writing Quality: Direct, concrete descriptions without purple prose
  • Natural Dialogue: Speech patterns feel authentic, not performative
  • Emotional Range: Handles vulnerability, humor, and character growth
  • Structural Variety: Avoids formulaic response patterns
  • Show Don't Tell: Trusts the reader, doesn't over-explain emotions

Comparison to Prima-24B

Trouper-12B and Prima-24B are trained on identical data but offer different trade-offs:

Aspect Trouper-12B Prima-24B
Prose Style Direct and concrete Slightly more elaborate
AI Slop Minimal Moderate (some patterns)
Reliability Good (template-sensitive) Excellent
Long Context Good (12B) Better (24B)
Inference Speed Faster (12B) Slower (24B)
Setup Difficulty Moderate (template critical) Easy
Action RP Good Excellent
Emotional RP Excellent Good

Choose Trouper-12B if: You want best prose quality, natural dialogue, and don't mind template setup
Choose Prima-24B if: You want reliability, long context, or action-oriented RP

Comparison to TinyRP-12B

This model addresses several issues found in my previous TinyRP-12B release:

Aspect TinyRP-12B Trouper-12B
Formulaic patterns Yes (after 20+ turns) No
Character stagnation Yes No - characters evolve
Opening variety Repetitive Varied
Training data Original dataset Custom "Actors" dataset
Long conversations Degrades Maintains quality

Known Limitations

  • Template Sensitivity: Without Mistral-V3-Tekken, may generate meta-narration or continue past appropriate stopping points
  • Occasional Meta-Breaks: Rare instances of stepping outside character (regenerate if needed)
  • Context Window: While good for 15-20+ turns, may be outperformed by larger models at 50+ turns. please let me know how it works for you!

Got Feedback?

Issues, questions, or feedback welcome! Particularly interested in:

  • Long conversation quality (20+ turns)
  • Template compatibility findings
  • Comparison with other RP models

Feel free to make a post in the Community tab here!

Downloads last month
273
Safetensors
Model size
12B params
Tensor type
F16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for DarwinAnim8or/Trouper-12B

Finetuned
(77)
this model
Quantizations
7 models