Models
GitHub
Discord
Docs
Cloud
Sign in
Download
Models
Download
GitHub
Discord
Docs
Cloud
Sign in
kaizu
/
bn_chat_2
:latest
14
Downloads
Updated
1 year ago
Fine-tuned version of llama2-v0.1-instruct from BanglaLLM in huggingface. Quantized to 4bit -> q4_k_m using llama.cpp. Trained on 2 * T4.
Fine-tuned version of llama2-v0.1-instruct from BanglaLLM in huggingface. Quantized to 4bit -> q4_k_m using llama.cpp. Trained on 2 * T4.
Cancel
Updated 1 year ago
1 year ago
fe73a4b5008c · 4.2GB ·
model
arch
llama
·
parameters
6.89B
·
quantization
Q4_K_M
4.2GB
template
[INST]<<SYS>>{{ .System }}<</SYS>> [INST]### Instruction: {{ .Prompt }}[/INST] ### Input: ### Respon
129B
system
নীচে একটি কাজের নির্দেশনা রয়েছে, এর স
335B
params
{ "num_ctx": 4096, "num_predict": 128, "repeat_penalty": 1.5, "stop": [ "[IN
174B
Readme
No readme
Write
Preview
Paste, drop or click to upload images (.png, .jpeg, .jpg, .svg, .gif)