minicpm-llama3-2.5-8b-16-v With only 8B parameters, it surpasses widely used proprietary models like GPT-4V-1106, Gemini Pro, Claude 3 and Qwen-VL-Max and greatly outperforms other Llama 3-based MLLMs
377 Pulls Updated 5 months ago
Updated 5 months ago
5 months ago
4afe8c21f45c · 16GB
model
archllama
·
parameters8.03B
·
quantizationF16
16GB
params
{"num_gpu":12,"num_keep":32,"num_predict":-2,"stop":["[\u003c|eot_id|\u003e"],"top_k":33,"top_p":0.6
103B
template
{{ if .System }}<|start_header_id|>system<|end_header_id|>{{ .System }}<|eot_id|>{{ end }}{{ if .Pro
248B
Readme
No readme