latest
5.1GB
medical assistant, responds to medical inquiries
7B
64 Pulls Updated 4 months ago
Updated 4 months ago
4 months ago
0e1793d9d93b · 5.1GB
model
archllama
·
parameters7.24B
·
quantizationQ5_K_M
5.1GB
system
You are a helpful assistant.
28B
params
{"stop":["[INST]","[/INST]"]}
30B
template
[INST] {{ .System }} {{ .Prompt }} [/INST]
42B
Readme
model takes up 5764 MB of GPU memory
GPU: Nvidia A10G
+---------------------------------------------------------------------------------------+
| NVIDIA-SMI 535.54.03 Driver Version: 535.54.03 CUDA Version: 12.2 |
|-----------------------------------------+----------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+======================+======================|
| 0 NVIDIA A10G On | 00000000:00:1E.0 Off | 0 |
| 0% 28C P0 57W / 300W | 5772MiB / 23028MiB | 0% Default |
| | | N/A |
+-----------------------------------------+----------------------+----------------------+
+---------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=======================================================================================|
| 0 N/A N/A 136397 C ...unners/cuda_v11/ollama_llama_server 5764MiB |
+---------------------------------------------------------------------------------------+