Roleplaying focused MoE Mistral model.

13B

173 Pulls Updated 6 weeks ago

6 weeks ago

87fc1d305675 · 7.8GB

model
llama
·
12.9B
·
Q4_K_M

Readme

Roleplaying focused MoE Mistral model.

One expert is a merge of mostly RP models, the other is a merge of mostly storywriting models. So it should be good at both. The base model is SanjiWatsuki/Kunoichi-DPO-v2-7B.

Expert 1 is a merge of LimaRP, Limamono, Noromaid 0.4 DPO and good-robot.
Expert 2 is a merge of Erebus, Holodeck, Dans-AdventurousWinds-Mk2, Opus, Ashhwriter and good-robot.

### Instruction:
{system prompt}

### Input:
User: {prompt}

### Response:
Character:

huggingface