A merge of WestLake-7B-v2 model with a good writing style and lots of details.
398 Pulls Updated 8 months ago
Updated 8 months ago
8 months ago
e1bcc8c2b80a · 11GB
model
archllama
·
parameters10.7B
·
quantizationQ8_0
11GB
params
{"stop":["USER:","ASSISTANT:"]}
32B
template
{{ .System }}
USER: {{ .Prompt }}
ASSISTANT:
48B
license
Apache License
Version 2.0, January
12kB
Readme
A merge of WestLake-7B-v2 model with a good writing style and lots of details. Performs very well for its size. Ranks above miqu-1-120b and goliath-120b. I strongly recommend the highest quantization you can run. See tags and download the version that suits you best.
Feel free to contact me if you have problems
Reddit: /u/spooknik | Discord: .spooknik