51 Downloads Updated 3 months ago
This is a quantized GGUF version of the powerful MaziyarPanahi/calme-3.2-instruct-78b model, optimized for use with Ollama.
| Metric | Value |
|---|---|
| Average | 52.02 |
| IFEval (0-Shot) | 80.63 |
| BBH (3-Shot) | 62.61 |
| MATH Lvl 5 (4-Shot) | 39.95 |
| GPQA (0-shot) | 20.36 |
| MuSR (0-shot) | 38.53 |
| MMLU-PRO (5-shot) | 70.03 |
ollama run richardyoung/calme-3.2:78b-q8
# Simple query
ollama run richardyoung/calme-3.2:78b-q8 "Explain quantum computing in simple terms"
# With system prompt
ollama run richardyoung/calme-3.2:78b-q8 --system "You are a helpful coding assistant" "Write a Python function to calculate fibonacci numbers"
You can customize the model behavior by creating a Modelfile:
FROM richardyoung/calme-3.2:78b-q8
# Adjust temperature for more creative responses
PARAMETER temperature 0.8
# Custom system prompt
SYSTEM "You are a knowledgeable AI assistant specialized in technical topics."
Then create your custom model:
ollama create my-calme -f ./Modelfile
This model uses the ChatML format:
<|im_start|>system
{System message}
<|im_end|>
<|im_start|>user
{User message}
<|im_end|>
<|im_start|>assistant
{Assistant response}
<|im_end|>
This model is subject to the Qwen License. Please review the license terms before use.
This GGUF conversion is provided for easier use with Ollama. All credit for the model development goes to MaziyarPanahi and the Qwen team.