kaizu/
bn_chat:latest

28 1 year ago

Fine-tuned version of llama2-v0.1-instruct from BanglaLLM in huggingface. Quantized to 4bit -> q4_k_m using llama.cpp.

672838002d45 · 148B
{
"num_predict": 128,
"repeat_penalty": 1.5,
"stop": [
"[INST]",
"[/INST]",
"<<SYS>>",
"<</SYS>>"
],
"temperature": 0.7
}