latest
9.0GB
Modelfile of orcamaid-13b.Q5_K_S.gguf from https://huggingface.co/TheBloke/OrcaMaid-13B-GGUF
13B
44 Pulls Updated 9 months ago
Updated 9 months ago
9 months ago
b8ded3a31488 · 9.0GB
model
archllama
·
parameters13.0B
·
quantizationQ5_K_S
9.0GB
template
{{ .System }}
### Instruction:
{{ .Prompt }}
### Response:
60B
params
{"num_ctx":4096,"stop":["</s>"]}
43B
system
You are OrcaMaid, an AI assistant. You carefully follow instructions.
70B
Readme
orcamaid-13b
Modelfile of orcamaid-13b.Q5_K_S.gguf from https://huggingface.co/TheBloke/OrcaMaid-13B-GGUF
Details:
- Name: orcamaid-13b.Q5_K_S.gguf
- Quant method: Q5_K_S
- Bits: 5
- Size: 8.97 GB
- Max RAM required: 11.47 GB
- Use case: large, low quality loss - recommended
Modelfile
FROM orcamaid-13b.Q5_K_S.gguf
TEMPLATE """{{ .System }} Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n### Instruction:\n{{ .Prompt }}\n\n### Response:"""
SYSTEM """You are OrcaMaid, an AI assistant. You carefully follow instructions."""
PARAMETER num_ctx 4096
PARAMETER stop "</s>"