168 5 months ago

kakaocorp/kanana-1.5-8b-instruct-2505 모델의 GGUF 버전

tools

5 months ago

dd34f3a73579 · 4.9GB ·

llama
·
8.03B
·
Q4_K_M
<|start_header_id|>system<|end_header_id|> {{ if .System }}{{ .System }} {{- end }} {{- if .Tools }}

Readme

카카오의 kakaocorp/kanana-1.5-8b-instruct-2505

허깅페이스에서 gguf-my-repo로 gguf 파일로 변환

(https://huggingface.co/spaces/ggml-org/gguf-my-repo)
양자화 (Q4_K_M)

기본 모델인 llama3.2 의 모델 파일 추출해서 사용함

(ollama show llama3.2 –modelfile >> Modefile)

FROM kanana-1.5-8b-instruct-2505-q4_k_m.gguf

TEMPLATE “”“<|start_header_id|>system<|end_header_id|>

{{ if .System }}{{ .System }} {{- end }} {{- if .Tools }}When you receive a tool call response, use the output to format an answer to the orginal user question.

You are a helpful assistant with tool calling capabilities. {{- end }}<|eotid|> {{- range \(i, \) := .Messages }} {{- \(last := eq (len (slice \).Messages \(i)) 1 }} {{- if eq .Role "user" }}<|start_header_id|>user<|end_header_id|> {{- if and \).Tools $last }}

Given the following functions, please respond with a JSON for a function call with its proper arguments that best answers the given prompt.

Respond in the format {“name”: function name, “parameters”: dictionary of argument name and its value}. Do not use variables.

{{ range $.Tools }} {{- . }} {{ end }} {{ .Content }}<|eot_id|> {{- else }}

{{ .Content }}<|eot_id|> {{- end }}{{ if $last }}<|start_header_id|>assistant<|end_header_id|>

{{ end }} {{- else if eq .Role “assistant” }}<|start_header_id|>assistant<|end_header_id|> {{- if .ToolCalls }} {{ range .ToolCalls }} {“name”: “{{ .Function.Name }}”, “parameters”: {{ .Function.Arguments }}}{{ end }} {{- else }}

{{ .Content }} {{- end }}{{ if not $last }}<|eot_id|>{{ end }} {{- else if eq .Role “tool” }}<|start_header_id|>ipython<|end_header_id|>

{{ .Content }}<|eot_id|>{{ if $last }}<|start_header_id|>assistant<|end_header_id|>

{{ end }} {{- end }} {{- end }}“”” PARAMETER stop <|start_header_id|> PARAMETER stop <|end_header_id|> PARAMETER stop <|eot_id|>