70K Downloads Updated 1 year ago
Updated 1 year ago
1 year ago
6f2d4b51d4ee · 43GB
Stable Beluga is based on Llama 2 and then fine-tuned on an Orca-style dataset. It is available in 7B, 13B, and 70B parameter sizes. It was created by Stability AI.
The model used in the example below is the Stable Beluga model, with 7b parameters, which is a general-use model.
ollama serve
)curl -X POST http://localhost:11434/api/generate -d '{
"model": "stable-beluga",
"prompt":"Explain the process of how a refrigerator works to keep the contents inside cold."
}'
ollama run stable-beluga
Note: The ollama run
command performs an ollama pull
if the model is not already downloaded. To download the model without running it, use ollama pull stable-beluga
If you run into issues with higher quantization levels, try using the q4 model or shut down any other programs that are using a lot of memory.
By default, Ollama uses 4-bit quantization. To try other quantization levels, please try the other tags. The number after the q represents the number of bits used for quantization (i.e. q4 means 4-bit quantization). The higher the number, the more accurate the model is, but the slower it runs, and the more memory it requires.
Aliases |
---|
latest, 7b, 7b-q4_0 |
13b, 13b-q4_0 |
70b, 70b-q4_0 |
Stable Beluga source on Ollama
7b parameters original source: Stability AI
13b parameters original source: Stability AI
70b parameters original source: Stability AI