llava:7b-v1.5-q3_K_L

6.9M 1 year ago

๐ŸŒ‹ LLaVA is a novel end-to-end trained large multimodal model that combines a vision encoder and Vicuna for general-purpose visual and language understanding. Updated to version 1.6.

vision 7b 13b 34b
7215dae26124 ยท 33B
{
"stop": [
"USER:",
"ASSSISTANT:"
]
}