The French General LLM
130 Pulls Updated 3 months ago
Updated 3 months ago
3 months ago
014b768ab174 · 2.3GB
model
archllama
·
parameters3.82B
·
quantizationQ4_K_M
2.3GB
template
{{ if .System }}<|system|>
{{ .System }}<|end|>
{{ end }}{{ if .Prompt }}<|user|>
{{ .Prompt }}<|end|>
{{ end }}<|assistant|>
{{ .Response }}<|end|>
155B
params
{"num_keep":4,"stop":["<|user|>","<|assistant|>","<|system|>","<|end|>","<|endoftext|>","###","<|fin|>"]}
166B
Readme
French-Alpaca is a general SLM in French
3.82B params
4k tokens of window context
Based on microsoft/Phi-3-mini-4k-instruct
Fine-tuned from the original French-Alpaca-dataset 110K instructions, entirely generated with OpenAI GPT-3.5-turbo.
The fine-tuning method is inspired from https://crfm.stanford.edu/2023/03/13/alpaca.html
HuggingFace model card : https://huggingface.co/jpacifico/French-Alpaca-Phi-3-mini-4k-instruct-v1.0-GGUF
Developed by: Jonathan Pacifico, 2024
License: MIT