Webvar
voyage-3 Embedding Model - logo

voyage-3 Embedding Model

Text embedding model optimized for general-purpose (including multilingual) retrieval/search and AI applications. 32K context length.
awsPurchase this listing from Webvar in AWS Marketplace using your AWS account. In AWS Marketplace, you can quickly launch pre-configured software with just a few clicks. AWS handles billing and payments, and charges on your AWS bill.

About

Text embedding models are neural networks that transform texts into numerical vectors. They are a crucial building block for semantic search/retrieval systems and retrieval-augmented generation (RAG) and are responsible for the retrieval quality. voyage-3 is a general-purpose embedding that: [1] outperforms OpenAI v3 large across all eight evaluated domains (tech, code, web, law, finance, multilingual, conservation, and long-context) by 7.55% on average, [2] has a 3-4x smaller embedding dimension (1024) compared to OpenAI (3072) and E5 Mistral (4096), resulting in 3-4x lower vectorDB costs, and [3] supports a 32K-token context length, compared to OpenAI (8K) and Cohere (512). Latency is 75 ms for a single query with at most 200 tokens, and throughput is 57M tokens per hour at $0.06 per 1M tokens on an ml.g6.xlarge. Learn more about voyage-3 here: https://blog.voyageai.com/2024/09/18/voyage-3/

Related Products

How it works?

Search

Search 25000+ products and services vetted by AWS.

Request private offer

Our team will send you an offer link to view.

Purchase

Accept the offer in your AWS account, and start using the software.

Manage

All your transactions will be consolidated into one bill in AWS.

Create Your Marketplace with Webvar!

Launch your marketplace effortlessly with our solutions. Optimize sales processes and expand your reach with our platform.