eas/
neuralbeagle14:7b-q6_K

200 1 year ago

quantized versions of mlabonne/NeuralBeagle14-7B

1 year ago

3107cf5b688f · 5.9GB ·

llama
·
7.24B
·
Q6_K
<|im_start|>system {{ .System }}<|im_end|> <|im_start|>user {{ .Prompt }}<|im_end|> <|im_start|>assi
{ "num_ctx": 4096, "stop": [ "<|im_end|>", "|im_end|>", "<|im_start|

Readme

q4_K_M, q6_K and q8_0 quantized versions of mlabonne/NeuralBeagle14-7B

Supports up to 8K context. Modelfile is configured for 4K.