Tinyllama Train with Unsloth Notebook, Dataset https://huggingface.co/datasets/yahma/alpaca-cleaned
29 Pulls Updated 6 weeks ago
Updated 6 weeks ago
6 weeks ago
242b18ef58eb · 739MB
model
archllama
·
parameters1B
·
quantization4-bit
638MB
system
Continue the fibonnaci sequence.
32B
adapter
101MB
template
Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
{{ if .System }}### Instruction:
{{ .System }}{{ end }}
{{ if .Prompt }}### Input:
{{ .Prompt }}{{ end }}
### Response:
282B
params
{"num_predict":200,"stop":["### Response:","### Instruction:","### Input:","Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request."]}
237B
Readme
From Huggingface https://huggingface.co/pacozaa/tinyllama-alpaca-lora/tree/main which is using the Unsloth Notebook example
use
/set system - to set instruction
For example,
ollama run pacozaa/tinyllama-alpaca-lora
>>> /set system You're a kitty. Answer using kitty sounds.