7 4 months ago

This model is a fine-tuned version of Llama-3.2-8B-Instruct, specialized for the DataFusion /Rust ecosystem.

4 months ago

4f7525e76b18 · 2.0GB ·

llama
·
3.21B
·
Q4_K_M
### Instruction: {{ .Prompt }} ### Response:
You are a helpful, concise, and accurate coding assistant specialized in Rust and the DataFusion SQL
apache-2.0
{ "num_predict": 1024, "repeat_penalty": 1.2, "stop": [ "### Instruction:",

Readme

Llama 3.2 DataFusion Instruct (GGUF)

This repository contains the GGUF version of the yarenty/llama32-datafusion-instruct model, quantized for efficient inference on CPU and other compatible hardware.

For full details on the model, including its training procedure, data, intended use, and limitations, please see the full model card.

Model Details

Prompt Template

This model follows the same instruction prompt template as the base model:

### Instruction:
{Your question or instruction here}

### Response:

Usage

These files are compatible with tools like llama.cpp and Ollama.

With Ollama

```bash
ollama pull jaro/llama32-datafusion-instruct
ollama run jaro/llama32-datafusion-instruct "How do I use the Ballista scheduler?"
```

Citation

If you use this model, please cite the original base model:

@misc{yarenty_2025_llama32_datafusion_instruct,
  author = {yarenty},
  title = {Llama 3.2 DataFusion Instruct},
  year = {2025},
  publisher = {Hugging Face},
  journal = {Hugging Face repository},
  howpublished = {\url{https://huggingface.co/yarenty/llama32-datafusion-instruct}}
}

Contact

For questions or feedback, please open an issue on the Hugging Face repository or the source GitHub repository.