A merge of WestLake-7B-v2 model with a good writing style and lots of details.
398 Pulls Updated 8 months ago
Updated 8 months ago
8 months ago
9a08a37e3227 · 7.4GB
model
archllama
·
parameters10.7B
·
quantizationQ5_K_S
7.4GB
params
{"stop":["USER:","ASSISTANT:"]}
32B
template
{{ .System }}
USER: {{ .Prompt }}
ASSISTANT:
48B
license
Apache License
Version 2.0, January
12kB
Readme
A merge of WestLake-7B-v2 model with a good writing style and lots of details. Performs very well for its size. Ranks above miqu-1-120b and goliath-120b. I strongly recommend the highest quantization you can run. See tags and download the version that suits you best.
Feel free to contact me if you have problems
Reddit: /u/spooknik | Discord: .spooknik