
Overview
Nomic Embed is the only truly open (open source, weights, and data) text embedder to beat OpenAI on both short and long context tasks.
Highlights
- Compared to similar models such as OpenAI's text-embedding-3-small, Nomic Embed outperforms on short context (MTEB 62.39 vs 62.26) and long context (LoCo 85.53 vs 82.40) benchmarks.
- Nomic embed supports binary embeddings, which can reduce the memory footprint of vector collections by several orders of magnitude.
- Nomic embed supports Matryoshka embeddings, allowing the user to have fine tuned control over the performance/disk space tradeoff.
Details
Unlock automation with AI agent solutions

Features and programs
Financing for AWS Marketplace purchases
Pricing
Dimension | Description | Cost/host/hour |
|---|---|---|
ml.p3.2xlarge Inference (Batch) Recommended | Model inference on the ml.p3.2xlarge instance type, batch mode | $1.50 |
ml.g5.xlarge Inference (Real-Time) Recommended | Model inference on the ml.g5.xlarge instance type, real-time mode | $2.50 |
ml.p2.xlarge Inference (Batch) | Model inference on the ml.p2.xlarge instance type, batch mode | $1.50 |
ml.g4dn.4xlarge Inference (Batch) | Model inference on the ml.g4dn.4xlarge instance type, batch mode | $2.60 |
ml.g4dn.16xlarge Inference (Batch) | Model inference on the ml.g4dn.16xlarge instance type, batch mode | $9.50 |
ml.p2.16xlarge Inference (Batch) | Model inference on the ml.p2.16xlarge instance type, batch mode | $1.50 |
ml.p3.16xlarge Inference (Batch) | Model inference on the ml.p3.16xlarge instance type, batch mode | $1.50 |
ml.g5.xlarge Inference (Batch) | Model inference on the ml.g5.xlarge instance type, batch mode | $1.60 |
ml.g5.8xlarge Inference (Batch) | Model inference on the ml.g5.8xlarge instance type, batch mode | $11.75 |
ml.g5.12xlarge Inference (Batch) | Model inference on the ml.g5.12xlarge instance type, batch mode | $17.87 |
Vendor refund policy
Please contact support@nomic.ai for inquires about refunds.
How can we make this page better?
Legal
Vendor terms and conditions
Content disclaimer
Delivery details
Amazon SageMaker model
An Amazon SageMaker model package is a pre-trained machine learning model ready to use without additional training. Use the model package to create a model on Amazon SageMaker for real-time inference or batch processing. Amazon SageMaker is a fully managed platform for building, training, and deploying machine learning models at scale.
Version release notes
task_type now accepted in request body
Additional details
Inputs
- Summary
Using our python client, you can simply pass this model a list of strings to embed.
If you are using Boto directly, you can hit the endpoint like
import boto3 data = { "texts": ["The quick brown fox jumps over the lazy dog"], "task_type": "search_document", } client = boto3.client("sagemaker-runtime", region_name='us-east-2') response = client.invoke_endpoint( EndpointName=ENDPOINT_NAME, Body=json.dumps(data), ContentType='application/json', )
- Input MIME type
- text/csv, application/json
Input data descriptions
The following table describes supported input data fields for real-time inference and batch transform.
Field name | Description | Constraints | Required |
|---|---|---|---|
csv | Input data should be a csv file without any column headers with each line containing a single text. | Type: FreeText | Yes |
Resources
Support
Vendor support
Please contact support@nomic.ai to inquire about support packages
AWS infrastructure support
AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.
Similar products

Customer reviews
Provide base layout and styling is also nice
Innovative and friendly technology