7 Downloads Updated 6 days ago
Updated 6 days ago
6 days ago
0a9ebafa4ede · 7.3GB ·
Atom V1 Preview 12B is a fine-tuned conversational AI model based on Google’s Gemma 3 12B Instruct architecture. This model is designed to function as a collaborative thought partner, specializing in exploratory dialogue, brainstorming, research assistance, and technical problem-solving while maintaining an approachable and engaging conversational style.
This 12B iteration of the Atom persona is the third release in Project Atom from VANTA Research, and is also our largest model to date.
Model Type: Multimodal Transformer (Text + Vision)
Base Model: google/gemma-3-12b-it
Training Method: Low-Rank Adaptation (LoRA) fine-tuning
License: Gemma Terms of Use
Developed By: VANTA Research
Language: English
The model employs a hybrid attention pattern with sliding window attention and periodic full attention layers (every 6th layer) for efficient long-context processing.
Atom-v1-preview-12b was fine-tuned using parameter-efficient LoRA adapters targeting attention and feedforward components. The training data consists of curated conversational examples emphasizing:
Training was conducted over 258 steps with careful monitoring to preserve the base model’s technical capabilities while introducing enhanced conversational characteristics.
This model is a research preview and should not be used for: - High-stakes decision-making without human oversight - Medical, legal, or financial advice - Generation of harmful, biased, or misleading content - Applications requiring guaranteed factual accuracy
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained(
"atom-v1-preview-12-hf",
torch_dtype="auto",
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("atom-v1-preview-12-hf")
messages = [
{"role": "user", "content": "What's your approach to explaining quantum entanglement?"}
]
inputs = tokenizer.apply_chat_template(
messages,
add_generation_prompt=True,
return_tensors="pt"
).to(model.device)
outputs = model.generate(
inputs,
max_new_tokens=512,
temperature=0.8,
top_p=0.9,
top_k=40,
do_sample=True
)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
Based on systematic evaluation across conversational dimensions:
The model demonstrates 85-90% alignment with design specifications across diverse prompt types, including identity awareness, technical discussion, creative output, empathetic support, and philosophical reasoning.
This model is designed to support exploration and learning, not to replace human judgment. Users should:
@misc{atom-v1-preview-12,
title={Atom-v1-preview-12: A Collaborative Thought Partner},
author={VANTA Research},
year={2025},
howpublished={https://huggingface.co/vanta-research/atom-v1-preview-12b}
}
Built on Google’s Gemma 3 12B Instruct architecture. Training infrastructure supported by Hugging Face Spaces, Transformers, PEFT, and llama.cpp quantization tools. Atom V1 12B was trained on NVIDIA’s L40S GPU