63 Downloads Updated 2 weeks ago
A compact yet powerful vision-language model from Hugging Face.
| Tag | Quantization | Size | Notes |
|---|---|---|---|
latest |
Q8_0 | ~2.4GB | Default |
q8 |
Q8_0 | ~2.4GB | Same as latest |
fp16 |
F16 | ~4.4GB | Full precision |
# Default (Q8)
ollama run ahmadwaqar/smolvlm2-2.2b-instruct "Describe this image" --images photo.jpg
# Explicit Q8
ollama run ahmadwaqar/smolvlm2-2.2b-instruct:q8 "Describe this image" --images photo.jpg
# FP16 (higher quality)
ollama run ahmadwaqar/smolvlm2-2.2b-instruct:fp16 "Describe this image" --images photo.jpg
from ollama import Client
client = Client(host='http://localhost:11434')
response = client.chat(
model='ahmadwaqar/smolvlm2-2.2b-instruct', # uses Q8 by default
messages=[{
'role': 'user',
'content': 'What do you see?',
'images': ['image.png']
}]
)
print(response['message']['content'])
| Property | Value |
|---|---|
| Parameters | 2.2B |
| Architecture | SmolLM2-1.7B + SigLIP |
| Context | 8K tokens |
| Variants | Q8_0 (default), FP16 |
| License | Apache 2.0 |