Cohere in Amazon Bedrock

Build enterprise AI applications that understand your business

Introducing Cohere’s Enterprise Foundation Models

Cohere's reranker model, Cohere Rerank 3.5, provides a powerful semantic boost to the search quality of any keyword or vector search system. In RAG use cases, reranking ensures that only the most relevant information is passed to the model. This provides better responses, reduced latency, and lower costs because the model processes less information.

Command R+ is Cohere's most advanced large language model, designed specifically for real-world enterprise applications. Command R+ balances efficiency and accuracy, empowering businesses to move beyond proof-of-concept and start utilizing AI in day-to-day operations. It supports 10 key business languages and excels at retrieval-augmented generation (RAG) use cases.

Command R is a powerful and versatile language model designed for businesses. It supports 10 languages and excels at long-context tasks, making it ideal for global enterprises. With a focus on efficiency and accuracy, Command R is optimized for RAG use cases. It adeptly handles text-generation tasks and is well-suited for full-scale AI implementation in enterprises.

Cohere Embed is a text embedding model that offers leading performance in 100+ languages. It translates text into vector representations which encode semantic meaning. Enterprises use this model to power search and retrieval systems. It is capable of outputting compressed embeddings (int8 and binary) to improve latency and reduce storage costs.

Benefits

With a context window of up to 128K tokens, the Command R models understand and generate responses within a broad context, making them ideal for complex workflows with large document ingestion, relevant citations with advanced retrieval, and tool use.
The Command R models have the capability for multilingual generation across 10 key business languages including: English, French, Spanish, Italian, German, Portuguese, Japanese, Korean, Arabic, and Chinese.
Command R+ supports multi-step tool use which allows the model to combine multiple tools over multiple steps to accomplish difficult tasks. The model can even correct itself when it tries to use a tool and fails, enabling the model to make multiple attempts at accomplishing the task and increasing the overall success rate.
Command R models are designed to enhance productivity by seamlessly integrating generative AI capabilities into everyday apps and workflows. Businesses can now streamline their processes and improve overall efficiency, leading to better business outcomes. With Command R+, enterprises can unlock new possibilities and elevate employee and customer experiences.
Cohere instills robust data privacy measures, allowing customers to retain complete control over their data. From customization to model inputs and outputs, businesses can rest assured that their sensitive information remains secure and under their supervision.

Meet Cohere's Command FM

Command is a text generation model for business use cases.

Use cases

Craft your message with an AI assistant, so you can write more clear and succinct emails.

Capture key points from an email chain, financial report, or customer call recording.

Provide users with more relevant and personalized search results through semantic search, designed to match the user intent behind a query.

Ask questions and get answers from your company’s entire knowledge base - from your messaging platform, to your cloud storage provider and CRM. Answers come with citations so you can confirm accuracy.

Input a set of data and have your AI assistant provide you with takeaways.

Model version

Rerank 3.5

Enhances search accuracy by reranking keyword and vector results, ensuring only the most relevant content reaches the model—delivering better responses while reducing both latency and costs.

Max tokens: 4,096

Languages: English, Chinese, Korean, Hindi, Japanese, Spanish, German, French, Arabic, Russian, Portuguese, and more. 

Fine-tuning supported: No

Supported use cases: Search-heavy, document-heavy, and RAG scenarios (Example: searching for a hotel)

Read the blog

Command R+

Command R+ is Cohere's most powerful generative language model optimized for long-context tasks, such as retrieval-augmented generation (RAG) and multi-step tool use.


Max tokens: 128K

Languages: English, French, Spanish, Italian, German, Portuguese, Japanese, Korean, Arabic, and Chinese

Fine-tuning supported: No

Supported use cases: Text generation, text summarization, chat, knowledge assistants, Q&A, RAG.

Read the blog

Command R

Command R is Cohere's generative language model optimized for long-context tasks, such as retrieval-augmented generation (RAG) and tools, and large scale production workloads.

Max tokens: 128K

Languages: English, French, Spanish, Italian, German, Portuguese, Japanese, Korean, Arabic, and Chinese

Fine-tuning supported: No

Supported use cases: Text generation, text summarization, chat, knowledge assistants, Q&A, RAG.

Read the blog

Command

Command is Cohere’s generative large language model (LLM).

Max tokens: 4K

Languages: English

Fine-tuning supported: Yes

Supported use cases: Chat, text generation, text summarization.

Read the blog

Command Light

Command Light is a smaller version of Command, Cohere's generative LLM.

Max tokens: 4K

Languages: English

Fine-tuning supported: Yes

Supported use cases: Chat, text generation, text summarization.

Read the blog

Embed - English

Embed is Cohere's text representation, or embeddings, model.
This version supports English only.

Dimensions: 1024

Languages: English

Fine-tuning supported: No

Supported use cases: Semantic search, retrieval augmented generation (RAG), classification, clustering.

Read the blog

Embed - Multilingual

Embed is Cohere's text representation, or embeddings, model.
This version supports multiple languages.

Dimensions: 1024

Languages: Multilingual (100+ supported languages)

Fine-tuning supported: No

Supported use cases: Semantic search, retrieval-augmented generation (RAG), classification, clustering.

Read the blog