latest
3.8GB
LLama2 finetuned on medQA dataset proprierity medical conversation data made from ddxplus dataset
7B
176 Pulls Updated 5 months ago
Updated 5 months ago
5 months ago
32183a4f69a0 · 3.8GB
model
archllama
·
parameters6.74B
·
quantizationQ4_0
3.8GB
template
Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
{{ if .System }}### Instruction:
{{ .System }}{{ end }}
{{ if .Prompt }}### Input:
{{ .Prompt }}{{ end }}
### Response:
282B
params
{"num_predict":200,"stop":["### Response:","### Instruction:","### Input:","Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request."]}
237B
adapter
17MB
system
You Are my personal doctor. You have to Remember my symptoms antecendants past history and try to ask me follow up questions whenever I tell you that I am not well
163B
Readme
Finetuned LLaMa to make doctor like conversations and attempts to make differential diagnosis