
Overview
Text embedding models are neural networks that transform texts into numerical vectors. They are a crucial building block for semantic search/retrieval systems and retrieval-augmented generation (RAG) and are responsible for the retrieval quality. voyage-law-2 is an embedding model optimized for retrieving legal texts. It excels in AI applications such as semantic case retrieval, legal question answering, and legal AI assistants. On 8 legal retrieval tasks, voyage-law-2 shows a 5.62% improvement over alternatives, including OpenAI v3 large, Cohere English v3, and E5 Mistral. It also enhances general-purpose corpora and long-context retrieval tasks, exceeding OpenAI v3 large by over 15% on average. Latency is 90 ms for a single query with at most 100 tokens, and throughput is 12.6M tokens per hour at $0.22 per 1M tokens on an ml.g6.xlarge. Learn more about voyage-law-2 here: https://blog.voyageai.com/2024/04/15/domain-specific-embeddings-and-retrieval-legal-edition-voyage-law-2/Â
Highlights
- Optimized for law retrieval. On 8 legal retrieval tasks, voyage-law-2 has a significant 5.62% improvement over any alternatives, including OpenAI v3 large, Cohere English v3 and E5 Mistral. Tops the [MTEB leaderboard for legal retrieval](https://huggingface.co/spaces/mteb/leaderboard?task=retrieval&language=law).
- Consistent enhancements across general-purpose corpora and long-context retrieval tasks, exceeding OpenAI v3 large on average by over 15%.
- 16K token context length, well-suited for applications on long documents. Latency is 90 ms for a single query with at most 100 tokens. 12.6M tokens per hour at $0.22 per 1M tokens on an ml.g6.xlarge.
Details
Unlock automation with AI agent solutions
