Updated 9 months ago
9 months ago
014b768ab174 · 2.3GB
model
archllama
·
parameters3.82B
·
quantizationQ4_K_M
2.3GB
template
{{ if .System }}<|system|>
{{ .System }}<|end|>
{{ end }}{{ if .Prompt }}<|user|>
{{ .Prompt }}<|
155B
params
{
"num_keep": 4,
"stop": [
"<|user|>",
"<|assistant|>",
"<|system|>"
166B
Readme
French-Alpaca is a general SLM in French 3.82B params 4k tokens of window context
Based on microsoft/Phi-3-mini-4k-instruct Fine-tuned from the original French-Alpaca-dataset 110K instructions, entirely generated with OpenAI GPT-3.5-turbo. The fine-tuning method is inspired from https://crfm.stanford.edu/2023/03/13/alpaca.html
HuggingFace model card : https://huggingface.co/jpacifico/French-Alpaca-Phi-3-mini-4k-instruct-v1.0-GGUF Developed by: Jonathan Pacifico, 2024 License: MIT