56K Downloads Updated 1 year ago
Yarn Mistral is a model based on Mistral that extends its context size up to 128k context. It is developed by Nous Research by implementing the YaRN method to further train the model to support larger context windows.
64k context size:
ollama run yarn-mistral
128k context size:
ollama run yarn-mistral:7b-128k
Example:
curl -X POST http://localhost:11434/api/generate -d '{
"model": "yarn-mistral:7b-128k",
"prompt":"Here is a story about llamas eating grass"
}'
YaRN: Efficient Context Window Extension of Large Language Models