
voyage-3 Embedding Model

About
Text embedding models are neural networks that transform texts into numerical vectors. They are a crucial building block for semantic search/retrieval systems and retrieval-augmented generation (RAG) and are responsible for the retrieval quality. voyage-3 is a general-purpose embedding that: [1] outperforms OpenAI v3 large across all eight evaluated domains (tech, code, web, law, finance, multilingual, conservation, and long-context) by 7.55% on average, [2] has a 3-4x smaller embedding dimension (1024) compared to OpenAI (3072) and E5 Mistral (4096), resulting in 3-4x lower vectorDB costs, and [3] supports a 32K-token context length, compared to OpenAI (8K) and Cohere (512). Latency is 75 ms for a single query with at most 200 tokens, and throughput is 57M tokens per hour at $0.06 per 1M tokens on an ml.g6.xlarge. Learn more about voyage-3 here: https://blog.voyageai.com/2024/09/18/voyage-3/
Related Products
show moreHow it works?
Search
Search 25000+ products and services vetted by AWS.
Request private offer
Our team will send you an offer link to view.
Purchase
Accept the offer in your AWS account, and start using the software.
Manage
All your transactions will be consolidated into one bill in AWS.
Create Your Marketplace with Webvar!
Launch your marketplace effortlessly with our solutions. Optimize sales processes and expand your reach with our platform.