Models
GitHub
Discord
Docs
Cloud
Sign in
Download
Models
Download
GitHub
Discord
Docs
Cloud
Sign in
kaizu
/
bn_chat
:latest
28
Downloads
Updated
1 year ago
Fine-tuned version of llama2-v0.1-instruct from BanglaLLM in huggingface. Quantized to 4bit -> q4_k_m using llama.cpp.
Fine-tuned version of llama2-v0.1-instruct from BanglaLLM in huggingface. Quantized to 4bit -> q4_k_m using llama.cpp.
Cancel
Updated 1 year ago
1 year ago
167e7d113012 · 4.2GB ·
model
arch
llama
·
parameters
6.89B
·
quantization
Q4_K_M
4.2GB
template
[INST]<<SYS>>{{ .System }}<</SYS>> [INST]### Instruction: {{ .Prompt }}[/INST] ### Input: ### Respon
129B
params
{ "num_predict": 128, "repeat_penalty": 1.5, "stop": [ "[INST]", "[/INST
148B
system
নীচে একটি কাজের নির্দেশনা রয়েছে, এর স
335B
Readme
No readme
Write
Preview
Paste, drop or click to upload images (.png, .jpeg, .jpg, .svg, .gif)