2 Downloads Updated yesterday
Updated yesterday
yesterday
0d9199b53959 · 4.5GB ·
Atom-Astronomy-7B is a specialized large language model fine-tuned for astronomy and astrophysics research. Built on the OLMo-3-7B-Instruct foundation, this model combines deep domain expertise with efficient inference, delivering graduate-level astronomical knowledge with 2.2x faster response times compared to competing models.
The model was trained on a carefully curated dataset comprising:
Astronomy Fundamentals
Stellar Physics
Cosmology
High-Energy Astrophysics
Galactic and Extragalactic Astronomy
Computational and Observational Methods
Specialized Topics
A comprehensive 10-question benchmark covering advanced topics including: - Eddington luminosity and super-Eddington accretion - Tolman-Oppenheimer-Volkoff equation derivations - Cosmological inflation and CMB physics - Relativistic beaming in gamma-ray bursts - Stellar nucleosynthesis (pp-chain and CNO cycle) - Cosmological recombination and Saha equation - Black hole orbital dynamics and ISCO calculations - Penrose process and Blandford-Znajek mechanism - Type Ia supernovae as standard candles - Hawking radiation and black hole thermodynamics
Results:
| Model | Avg Response Time | Total Time | Avg Words | Equation Usage | Calculation Rate |
|---|---|---|---|---|---|
| Atom-Astronomy-7B | 75.2s | 12.5 min | 2,032 | 100% | 100% |
| OLMo-3-7B-Instruct | 125.2s | 20.9 min | 3,396 | 100% | 100% |
| Qwen3-8B | 168.0s | 28.0 min | 5,277 | 100% | 100% |
Key Findings: - 2.23x faster than Qwen3-8B - 1.67x faster than base OLMo-3-7B - Maintains 100% technical accuracy with equations and calculations - Delivers concise, focused responses without sacrificing depth - 40-55% reduction in total processing time for complex queries
Status: Evaluation in progress
This model is currently undergoing comprehensive evaluation on the AstroBench_MCQ_v1_Public dataset, a professional-grade multiple-choice question benchmark derived from the Annual Review of Astronomy and Astrophysics. The dataset contains 3,846 expert-level questions covering the full breadth of modern astronomy research.
Preliminary Observations: - 90% answer extraction rate (18⁄20 in initial test) - 5.43s average response time per question - Maintains technical reasoning quality with proper elimination of incorrect options - Shows appropriate caution by not forcing answers when uncertain
Full results will be published upon completion of the comprehensive evaluation. The model card will be updated with detailed accuracy metrics and comparative analysis.
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
model_name = "vanta-research/atom-astronomy-7b"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype=torch.float16,
device_map="auto"
)
prompt = """Explain the Tolman-Oppenheimer-Volkoff equation and how it differs from
standard hydrostatic equilibrium. What does this tell us about neutron star structure?"""
messages = [
{"role": "system", "content": "You are Atom, a helpful AI assistant specialized in astronomy and astrophysics."},
{"role": "user", "content": prompt}
]
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to(model.device)
outputs = model.generate(inputs, max_new_tokens=1024, temperature=0.7, top_p=0.9)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
For efficient local deployment, quantized GGUF versions are available:
# Using Ollama
ollama create atom-astronomy:7b -f Modelfile
# Query the model
ollama run atom-astronomy:7b "Calculate the Schwarzschild radius for a 10 solar mass black hole"
VANTA Research
If you use Atom-Astronomy-7B in your research, please cite:
@misc{atom-astronomy-7b,
title={Atom-Astronomy-7B: A Specialized Language Model for Astronomy and Astrophysics},
author={VANTA Research},
year={2025},
publisher={Hugging Face},
howpublished={\url{https://huggingface.co/vanta-research/atom-astronomy-7b}}
}
Please also cite the base model:
@article{olmo3,
title={OLMo 3: Open Language Model},
author={Allen Institute for AI},
year={2024},
}
This model builds upon the excellent work of the Allen Institute for AI in developing the OLMo series of open language models. We thank the astronomy and astrophysics community for developing the open-source educational materials and research papers that informed our training data curation.
For questions, issues, or collaboration inquiries, please contact: - Email: hello@vantaresearch.xyz