765 3 months ago

tools

Models

View all →

Readme

Description:

The Qwen3-Embedding-0.6B-GGUF is a specialized text embedding model, part of the Qwen3 series, tailored for efficient and effective text embedding and reranking tasks.

Key Features & Use Cases:

  • Model Type: Primarily a Text Embedding model, also part of a series that includes Text Reranking models.
  • Size: A compact model with 0.6 billion parameters, making it suitable for scenarios where efficiency is prioritized. Other sizes (4B, 8B) are available in the series for different performance needs.
  • Multilingual Capability: Supports over 100 languages, including various programming languages, excelling in multilingual, cross-lingual, and code retrieval scenarios.
  • Context Length: Capable of processing texts up to 32,000 tokens.
  • Embedding Dimension: Offers flexible output dimensions, allowing users to define embeddings from 32 to 1024.
  • Instruction Awareness: Supports user-defined instructions to optimize performance for specific tasks, languages, or use cases. This can lead to a 1% to 5% improvement in performance for most downstream tasks, especially in retrieval scenarios.
  • Quantization: Available in q8_0 and f16 quantization formats.
  • Applications: Ideal for various text embedding and ranking tasks, including:
    • Text Retrieval
    • Code Retrieval
    • Text Classification
    • Text Clustering
    • Bitext Mining

Usage Notes:

  • When using the model, it’s recommended to append the <|endoftext|> token to your input context.
  • For optimal results, tailor your input instructions (instruct) to the specific task or language.

HuggingFace