480 1 month ago

An optimized version of Google's TranslateGemma-12B-it (Gemma 3) designed for high-fidelity translation. This build features hard-coded Temperature=0.1 and English Anchor support to eliminate output redundancy and maximize accuracy.

12b
ollama run rinex20/translategemma3:12b

Models

View all →

Readme

TranslateGemma-12B-it-Optimized (by rinex20)

This model is a highly refined version of Google’s TranslateGemma-12B-it, based on the state-of-the-art Gemma 3 architecture. It has been specifically optimized via a custom Modelfile to eliminate common local deployment issues such as instruction drift, output redundancy, and inconsistent terminology.

🚀 Key Features

  • Outperforms 27B: The 12B variant is empirically proven to exceed the translation quality of the Gemma 3 27B baseline on the MetricX benchmark.
  • English Instruction Anchoring: Optimized to respond to English-based translation triggers (e.g., To English:, To Japanese:), providing a 100% stable “translation mode” activation.
  • Auto-Language Detection: Robust multilingual awareness; the model automatically identifies the source language.
  • Zero-Filler Output: Forced “Deterministic Mode” to provide the translation result only, without conversational small talk or metadata.
  • Terminology Guard: Built-in system logic to protect technical terms (e.g., Kubernetes, Ollama, PyTorch) from being over-translated.

🛠 Quick Start (Ollama)

Run the model directly using the following command:

ollama run rinex20/translategemma3:12b

💡 Prompting Best Practices

For research-grade accuracy and consistency, use English Anchors as prefixes. This effectively “switches” the model into its specialized translation sub-network:

Task Recommended Input Format
Translate to English To English: [Your Text]
Translate to Japanese To Japanese: [Your Text]
Translate to Chinese To Chinese: [Your Text]
Auto-Detection Translate to [Target Language]: [Your Text]

⚙️ Model Configurations

The following parameters are hard-coded into this version to ensure high-fidelity performance:

  • Temperature: 0.1 – Minimizes randomness for factual, deterministic translations.
  • Top_P: 0.9 – Maintains focus on the most semantically relevant tokens.
  • Num_Ctx: 8192 – Optimized context window for translating technical documentation or long-form paragraphs.

📜 Credits & License