๐ LLaVA is a novel end-to-end trained large multimodal model that combines a vision encoder and Vicuna for general-purpose visual and language understanding. Updated to version 1.6.
vision
7b
13b
34b
2M Pulls Updated 10 months ago
9fb057c3f08a ยท 45B
{{ .System }}
USER: {{ .Prompt }}
ASSSISTANT: