A merge of WestLake-7B-v2 model with a good writing style and lots of details.
398 Pulls Updated 7 months ago
Updated 7 months ago
7 months ago
fea8646d4a3c · 8.8GB
model
archllama
·
parameters10.7B
·
quantizationQ6_K
8.8GB
params
{"stop":["USER:","ASSISTANT:"]}
32B
template
{{ .System }}
USER: {{ .Prompt }}
ASSISTANT:
48B
license
Apache License
Version 2.0, January
12kB
Readme
A merge of WestLake-7B-v2 model with a good writing style and lots of details. Performs very well for its size. Ranks above miqu-1-120b and goliath-120b. I strongly recommend the highest quantization you can run. See tags and download the version that suits you best.
Feel free to contact me if you have problems
Reddit: /u/spooknik | Discord: .spooknik