Open-source medical large language model adapted from Llama 2 to the medical domain.
7b
70b
41.6K Pulls Updated 13 months ago
Updated 13 months ago
13 months ago
ad11a6250f54 · 3.8GB
model
archllama
·
parameters6.74B
·
quantizationQ4_0
3.8GB
params
{
"num_ctx": 2048,
"stop": [
"<|im_start|>",
"<|im_end|>"
]
}
74B
template
<|im_start|>system
{{ .System }}<|im_end|>
<|im_start|>user
{{ .Prompt }}<|im_end|>
<|im_start|>assi
106B
system
A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful,
155B
Readme
Meditron is a large language model adapted from Llama 2 to the medical domain through training on a corpus of medical data, papers and guidelines. It outperforms Llama 2, GPT 3.5 and Flan-PaLM on many medical reasoning tasks.
Potential use cases include:
- Medical exam question answering
- Supporting differential diagnosis
- Disease information (symptoms, cause, treatment) query
- General health information query
Example prompts
What are the symptoms of the common cold?
What causes the seasonal flu?
What medication would be prescribed for a headache?
References
MEDITRON-70B: Scaling Medical Pretraining for Large Language Models