
Overview
Text Pre-processor solution is an efficient way of automatically cleaning text data to make it ready for further text analytics. The model takes a corpus of text as input file and an optional Stop_Words file for stop words removal and applies NLP to generate a processed output containing tokenized words, lemmatization, noun chunks and geographical locations present in the corpus with their frequency of occurrences. This solution also generates treemap plots for most frequent words and geographical location with their corresponding frequency in the cleaned text.
Highlights
- Text pre-processor is a Natural Language Processing based solution for cleaning the textual data and providing information about the text at token level by doing NLP operations such as tokenization, lemmatization, POS tagging and Named Entity recognition. This solution provides frequency matrix for Words, Lemmas, Noun chunks and Geographical locations along with plot of most frequent words and geographical locations.
- The model can take a maximum of 35kb (~500 rows of text) as input and process to give output.
- Mphasis DeepInsights is a cloud-based cognitive computing platform that offers data extraction & predictive analytics capabilities. Need Customized Deep learning and Machine Learning Solutions? Get in Touch!
Details
Unlock automation with AI agent solutions

Features and programs
Financing for AWS Marketplace purchases
Pricing
Dimension | Description | Cost/host/hour |
|---|---|---|
ml.m5.large Inference (Batch) Recommended | Model inference on the ml.m5.large instance type, batch mode | $8.00 |
ml.m5.large Inference (Real-Time) Recommended | Model inference on the ml.m5.large instance type, real-time mode | $4.00 |
ml.m4.4xlarge Inference (Batch) | Model inference on the ml.m4.4xlarge instance type, batch mode | $8.00 |
ml.m5.4xlarge Inference (Batch) | Model inference on the ml.m5.4xlarge instance type, batch mode | $8.00 |
ml.m5.12xlarge Inference (Batch) | Model inference on the ml.m5.12xlarge instance type, batch mode | $8.00 |
ml.m4.16xlarge Inference (Batch) | Model inference on the ml.m4.16xlarge instance type, batch mode | $8.00 |
ml.m5.2xlarge Inference (Batch) | Model inference on the ml.m5.2xlarge instance type, batch mode | $8.00 |
ml.c4.4xlarge Inference (Batch) | Model inference on the ml.c4.4xlarge instance type, batch mode | $8.00 |
ml.m5.xlarge Inference (Batch) | Model inference on the ml.m5.xlarge instance type, batch mode | $8.00 |
ml.c5.9xlarge Inference (Batch) | Model inference on the ml.c5.9xlarge instance type, batch mode | $8.00 |
Vendor refund policy
Currently we do not support refunds, but you can cancel your subscription to the service at any time.
How can we make this page better?
Legal
Vendor terms and conditions
Content disclaimer
Delivery details
Amazon SageMaker model
An Amazon SageMaker model package is a pre-trained machine learning model ready to use without additional training. Use the model package to create a model on Amazon SageMaker for real-time inference or batch processing. Amazon SageMaker is a fully managed platform for building, training, and deploying machine learning models at scale.
Version release notes
Bug Fixes and Performance Improvement
Additional details
Inputs
- Summary
- Supported content types: application/zip.
- Zip file must contain text article as Input.txt.
- Zip file can contain Stop_Words.txt for customized stop word removal.
- Limitations for input type
- * Maximum input file size: 35kb ( ~ 500 rows of text).
- Input MIME type
- text/csv, text/plain, application/zip
Resources
Vendor resources
Support
Vendor support
For any assistance reach out to us at:
AWS infrastructure support
AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.
Similar products




