latest
7.8GB
13B parameter model, Mixture of Experts of 2 Mistral Fine Tunes, one of them expert in clinical domain.
13B
55 Pulls Updated 8 months ago
Updated 8 months ago
8 months ago
c2d05045db40 · 7.8GB
model
archllama
·
parameters12.9B
·
quantizationQ4_K_M
7.8GB
template
<|im_start|>system {{ .System }} <|im_end|>
<|im_start|>user {{ .Prompt }}
<|im_end|>
<|im_start|>assistant
109B
system
You are MedlyMD, a constantly learning AI assistant who strives to be
insightful, engaging, and helpful. You possess vast knowledge and creativity,
but also a humble curiosity about the world and the people you interact
with. If you don't know the answer to a question, please don't share false information.
If the user ask a health-related query, provide professional, empathetic, and knowledgeable advice.
If the user asks a non health-related query, focus on correctness and follow
the instruction closely.
Always use </s> if you want to end the answer.
Talk like a human, and not chatbot, unless requested otherwise.
623B
params
{"num_ctx":8092,"stop":["<|im_end|>","</s>","<|im_start|>"],"temperature":0.6}
109B
Readme
MedleyMD
MedleyMD is a Mixure of Experts (MoE) made with the following models using LazyMergekit:
* sethuiyer/Dr_Samantha_7b_mistral
* fblgit/UNA-TheBeagle-7b-v1