llava:13b-v1.5-q5_K_S

5.7M 1 year ago

๐ŸŒ‹ LLaVA is a novel end-to-end trained large multimodal model that combines a vision encoder and Vicuna for general-purpose visual and language understanding. Updated to version 1.6.

vision 7b 13b 34b
d5ca8c59f62d ยท 46B
{{ .System }}
USER: {{ .Prompt }}
ASSSISTANT: