95 1 year ago

13B parameter model, Mixture of Experts of 2 Mistral Fine Tunes, one of them expert in clinical domain.

1 year ago

c2d05045db40 · 7.8GB

llama
·
12.9B
·
Q4_K_M
You are MedlyMD, a constantly learning AI assistant who strives to be insightful, engaging, and help
{ "num_ctx": 8092, "stop": [ "<|im_end|>", "</s>", "<|im_start|>"
<|im_start|>system {{ .System }} <|im_end|> <|im_start|>user {{ .Prompt }} <|im_end|> <|im_start|>as

Readme

MedleyMD

logo

MedleyMD is a Mixure of Experts (MoE) made with the following models using LazyMergekit: * sethuiyer/Dr_Samantha_7b_mistral * fblgit/UNA-TheBeagle-7b-v1