8b
4.9GB
Stheno-v3.2-Zeta I have done a test run with multiple variations of the models, merged back to its base at various weights, different training runs too, and this Sixth iteration is the one I like most.
8B
1 Pull Updated 2 days ago
Updated 2 days ago
2 days ago
a348dcd0fcac · 4.9GB
model
archllama
·
parameters8.03B
·
quantizationQ4_K_M
4.9GB
params
{"min_p":0.075,"repeat_penalty":1.1,"stop":["\\n\\n{{User}}","<|eot_id|>","<|end_of_text|>"],"temperature":1.12,"top_k":50}
144B
template
{{ if .System }}<|begin_of_text|><|start_header_id|>system<|end_header_id|>
{{ .System }}<|eot_id|>{{ end }}{{ if .Prompt }}<|start_header_id|>user<|end_header_id|>
{{ .Prompt }}<|eot_id|>{{ end }}<|start_header_id|>assistant<|end_header_id|>
{{ .Response }}<|eot_id|>
277B
license
https://spdx.org/licenses/CC-BY-NC-4.0
Created by Sao10k
https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2
Quants by Lewdiculous
https://huggingface.co/Lewdiculous/L3-8B-Stheno-v3.2-GGUF-IQ-Imatrix
203B
Readme
No readme