32K
Embedding
Active

Qwen Text Embedding v3

Built on Qwen3 foundations, it prioritizes long-context understanding and semantic accuracy for real-world applications.
Try it now
Testimonials

Our Clients' Voices

Qwen Text Embedding v3Techflow Logo - Techflow X Webflow Template

Qwen Text Embedding v3

Qwen3 Text Embedding V3 delivers state-of-the-art performance in multilingual text embeddings.

Overview

Qwen Text Embedding V3 represents a cutting-edge embedding model optimized for dense vector representations, excelling in semantic search, retrieval-augmented generation (RAG), and multilingual similarity tasks across 100+ languages. It delivers high-dimensional embeddings up to 4096 in length, with dynamic dimensionality reduction for efficiency, enabling precise capture of nuanced meanings in long texts and cross-lingual contexts.

Technical Specifications

  • Architecture: Transformer-based dual-encoder with asymmetric fine-tuning
  • Vector Dimensionality: 1024 (configurable down to 256 via projection layers)
  • Supported Tasks: Semantic similarity, document clustering, cross-lingual retrieval, reranking support
  • Training Data: Multilingual corpus spanning 30+ languages, enriched with technical, academic, and conversational domains

Performance Benchmarks

  • Semantic retrieval (MTEB & BEIR): Top-tier scores in passage ranking and asymmetric search
  • Multilingual tasks: Outperforms prior versions by 5-10% on C-MTEB leaderboard

Output Quality & Semantic Fidelity

Users report marked improvements in vector consistency across paraphrases, domain shifts, and query-document asymmetry. Embeddings exhibit reduced topic drift in iterative retrieval systems and stronger alignment with human-judged relevance rankings. The model excels in distinguishing subtle sentiment and intent variations, critical for customer support routing and compliance filtering.

Quality Improvements

  • Noise resilience: Robust to OCR errors, informal syntax, and mixed-language inputs
  • Temporal stability: Embedding drift minimized over time, ensuring index compatibility in production systems
  • Bias mitigation: Enhanced fairness controls reduce stereotypical associations in gender, profession, and geographic representations

API Pricing

  • $0.0735 / 1M tokens

New Features & Technical Upgrades

Qwen Text Embedding v3 introduces several architectural and training innovations to push the boundaries of dense retrieval.

Key Features

  • Multilingual Mastery: Handles 100+ languages in a single embedding space, outperforming priors in cross-lingual retrieval by leveraging Qwen3's reasoning backbone.​
  • Flexible Prompting: Distinct query and document prefixes boost retrieval accuracy without retraining, ideal for RAG pipelines.​
  • Variable Dimensions: Customizable output sizes from 32 to max dims reduce latency while preserving quality.​
  • Task Versatility: Optimized for embedding plus reranking, with SOTA scores in code retrieval and bitext mining

Code Sample

Comparison with Other Models

  • vs text-embedding-3-large: Qwen v3 matches or exceeds OpenAI’s large embedding model on non-English benchmarks while offering 5x lower cost and on-prem deployment options
  • vs Cohere Embed V3: Delivers superior performance in code and technical document embedding, with stronger support for Asian languages
  • vs Qwen Embedding v2: +6.1% average gain on retrieval tasks, 40% lower latency, and native support for 8K context (vs. 4K in v2)
Try it now

400+ AI Models

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

The Best Growth Choice
for Enterprise

Get API Key