
Overview
Jamba 1.5 Large is the first of its kind hybrid Mamba-Transformer architecture at a production grade level offering unmatched efficiency. With an unprecedented context window length (256K), it offers superior quality output for tasks needing large input context & low latency, at a competitive price point for its class.
Highlights
- With a 256K effective long context window, Jamba 1.5 models lead the NVIDIA RULER benchmark—the standard for measuring effective context windows across practical tasks, improving the output quality of key enterprise workflows, such as lengthy document summarization and multi-document analysis.
- 2.5X faster than leading models in its size class on long context and fastest across all context lengths.
- Supports function calling/tool use, structured output (JSON), and grounded generation with citation mode and documents API.
Details
Unlock automation with AI agent solutions

Features and programs
Financing for AWS Marketplace purchases
Pricing
Dimension | Description | Cost |
|---|---|---|
ml.m5.large Inference (Batch) Recommended | Model inference on the ml.m5.large instance type, batch mode | $0.00/host/hour |
inference.count.m.i.c Inference Pricing | inference.count.m.i.c Inference Pricing | $60.00/request |
Vendor refund policy
No refunds.
How can we make this page better?
Legal
Vendor terms and conditions
Content disclaimer
Delivery details
Amazon SageMaker model
An Amazon SageMaker model package is a pre-trained machine learning model ready to use without additional training. Use the model package to create a model on Amazon SageMaker for real-time inference or batch processing. Amazon SageMaker is a fully managed platform for building, training, and deploying machine learning models at scale.
Version release notes
Model release
Additional details
Inputs
- Summary
Jamba 1.5 Large is the first of its kind hybrid Mamba-Transformer architecture at a production grade level offering unmatched efficiency. With an unprecedented context window length (256K), it offers superior quality output for tasks needing large input context & low latency, at a competitive price point for its class.
- Input MIME type
- application/json
Input data descriptions
The following table describes supported input data fields for real-time inference and batch transform.
Field name | Description | Constraints | Required |
|---|---|---|---|
messages | A list of message objects for the conversation history, which can include system, user, assistant, and tool messages.
See https://docs.ai21.com/reference/jamba-15-api-ref#request-parameters for full description. | Type: FreeText | Yes |
temperature | Controls randomness. allowed range [0,2] | Default value: 1
Type: Continuous
Minimum: 0
Maximum: 2 | No |
top_p | Controls diversity via nucleus sampling. | Default value: 1
Type: Continuous
Minimum: 0
Maximum: 1 | No |
n | Number of completions to generate for each prompt. | Default value: 1
Type: Integer
Minimum: 1
Maximum: 16 | No |
stop | Sequences where the API will stop generating further tokens. Up to 4 sequences | Default value: []
Type: FreeText | No |
max_tokens | Maximum number of tokens to generate. | Default value: 4096
Type: Integer | No |
response_format | Setting to { "type": "json_object" } enables JSON mode, which guarantees the message the model generates is valid JSON | Default value: null
Type: FreeText | No |
tools | A list of tools the model may call. Currently, only functions are supported as a tool. Use this to provide a list of functions the model may generate JSON inputs for. A max of 128 functions are supported.
See https://docs.ai21.com/reference/jamba-15-api-ref#request-parameters for full description. | Default value: []
Type: FreeText | No |
documents | A list of relevant documents the model can ground its responses on, if the user explicitly says so in the prompt. Essentially acts as an extension to the prompt, with the ability to add metadata. each document is a dictionary.
See https://docs.ai21.com/reference/jamba-15-api-ref#request-parameters for full description. | Default value: []
Type: FreeText | No |
Resources
Vendor resources
Support
Vendor support
Talk to our experts: support-amazon@ai21.comÂ
AWS infrastructure support
AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.
Similar products
