Listing Thumbnail

    Linux Generative AI & AI/ML Server (Preconfigured LLM Stack)

     Info
    Deployed on AWS
    Preconfigured Amazon Linux 2023 AMI with LLM tools, Hugging Face, LangChain, RStudio, and JupyterLab. Build and scale Generative AI using AWS Bedrock, SageMaker, and OpenSearch.

    Overview

    Overview

    This Amazon Linux 2023 AMI comes fully pre-installed with essential tools for Generative AI, data science, and machine learning workloads. It is optimized for scalable cloud deployments, AWS-native AI integrations, and secure remote access, delivering a seamless out-of-the-box experience for researchers, developers, and enterprise teams.

    Key Features

    Remote Access: High-performance remote desktop access via Amazon NICE DCV. Web & Utility Tools: Google Chrome, Git, and AWS CLI preinstalled. Container-Ready Environment: Docker and Docker Compose configured for scalable deployments and microservices-based AI applications. Development Environments: Visual Studio Code, PyCharm Community Edition, JupyterLab, and RStudio for end-to-end development. Generative AI Frameworks: Hugging Face Transformers, LangChain, LlamaIndex, FAISS, and LoRA/PEFT for fine-tuning LLMs. Sample Workflows: RAG chatbot with Bedrock + OpenSearch, multimodal pipelines, and MONAI for medical imaging use cases. ML & Data Science Frameworks: PyTorch, TensorFlow, scikit-learn, PySpark, Dask, and Vowpal Wabbit. Productivity Tools: LibreOffice for document editing and reporting. AWS-Native Integrations: Bedrock SDK, SageMaker SDK, and OpenSearch connectors for building scalable Generative AI applications.

    Technical Details

    Operating System: Amazon Linux 2023 Remote Access: Amazon NICE DCV Browsers & Utilities: Google Chrome, Git, AWS CLI Programming Languages: Python 3.x, R

    IDEs & Authoring Tools: Visual Studio Code PyCharm Community Edition RStudio Desktop & Server

    Notebook Interfaces: Jupyter Notebook JupyterLab

    GenAI & Machine Learning Frameworks: Hugging Face Transformers, LangChain, LlamaIndex FAISS for vector search and RAG pipelines PyTorch, TensorFlow, scikit-learn PySpark, Dask, Vowpal Wabbit MONAI for medical imaging LoRA and PEFT for fine-tuning LLMs

    Environment & Containerization: Anaconda Docker Docker Compose

    Office Tools: LibreOffice Writer, Calc, Impress

    Highlights

    • Preconfigured Generative AI Environment on Amazon Linux 2023 with Hugging Face, LangChain, FAISS, and LoRA for LLM fine tuning.
    • AWS Native Integrations with Bedrock, SageMaker, and OpenSearch for building RAG chatbots, multimodal workflows, and scalable AI applications.
    • Developer Productivity with Amazon NICE DCV remote access, JupyterLab, RStudio, Visual Studio Code, and a complete data science and AI development stack.

    Details

    Delivery method

    Delivery option
    64-bit (x86) Amazon Machine Image (AMI)

    Latest version

    Operating system
    AmazonLinux 2023

    Deployed on AWS
    New

    Introducing multi-product solutions

    You can now purchase comprehensive solutions tailored to use cases and industries.

    Multi-product solutions

    Features and programs

    Financing for AWS Marketplace purchases

    AWS Marketplace now accepts line of credit payments through the PNC Vendor Finance program. This program is available to select AWS customers in the US, excluding NV, NC, ND, TN, & VT.
    Financing for AWS Marketplace purchases

    Pricing

    Linux Generative AI & AI/ML Server (Preconfigured LLM Stack)

     Info
    This product is available free of charge. Free subscriptions have no end date and may be canceled any time.
    Additional AWS infrastructure costs may apply. Use the AWS Pricing Calculator  to estimate your infrastructure costs.

    Vendor refund policy

    NA

    How can we make this page better?

    We'd like to hear your feedback and ideas on how to improve this page.
    We'd like to hear your feedback and ideas on how to improve this page.

    Legal

    Vendor terms and conditions

    Upon subscribing to this product, you must acknowledge and agree to the terms and conditions outlined in the vendor's End User License Agreement (EULA) .

    Content disclaimer

    Vendors are responsible for their product descriptions and other product content. AWS does not warrant that vendors' product descriptions or other product content are accurate, complete, reliable, current, or error-free.

    Usage information

     Info

    Delivery details

    64-bit (x86) Amazon Machine Image (AMI)

    Amazon Machine Image (AMI)

    An AMI is a virtual image that provides the information required to launch an instance. Amazon EC2 (Elastic Compute Cloud) instances are virtual servers on which you can run your applications and workloads, offering varying combinations of CPU, memory, storage, and networking resources. You can launch as many instances from as many different AMIs as you need.

    Version release notes

    NA

    Additional details

    Usage instructions

    "Quick Usage Summary

    1. Subscribe to the AWS Marketplace product and launch an instance.
    2. Use at least 200 GB for the EBS Volume and select g4dn.4xlarge for smooth performance.
    3. AI and ML libraries are installed in the conda environment named ""genai"". To use them: a. Open Anaconda Prompt. b. Run: conda activate genai c. Run: python --version d. Import packages such as torch, transformers, langchain, faiss, scikit-learn, pyspark, dask, vowpalwabbit, monai, peft.

    Connect via NICE DCV

    1. Open a browser and navigate to https://<your-public-dns-or-IP>:8443
    2. Log in using your Windows Administrator username and password.
    3. You will gain access to the Windows desktop in your browser. Note: Ensure that TCP port 8443 is allowed in the EC2 security group and Windows firewall.

    Development IDEs

    1. Launch Visual Studio Code, Visual Studio 2022, or PyCharm from the Start Menu.
    2. Create new files or open existing projects.
    3. Suitable for Python, R, .NET, and full-stack development.

    Anaconda

    1. Open Anaconda Navigator from the Start Menu to manage environments and packages.
    2. Alternatively, use Anaconda Prompt to run commands such as: conda list

    JupyterLab and Python

    1. Access JupyterLab from the desktop or Start Menu.
    2. Run Python or R notebooks and import preinstalled libraries.
    3. Activate the genai environment for AI and ML workflows.

    Machine Learning and AI Libraries

    The environment includes preinstalled libraries for data science, machine learning, and generative AI. Examples:

    1. Transformers version 4.56.2
    2. LangChain version 0.3.27
    3. LlamaIndex version 0.14.3
    4. FAISS version 1.9.0
    5. PyTorch version 2.5.1 with CUDA 12.1 (GPU enabled)
    6. scikit-learn version 1.7.2
    7. PySpark version 4.0.1
    8. Dask version 2025.9.1
    9. VowpalWabbit version 9.10.0
    10. MONAI version 1.5.1
    11. PEFT version 0.17.1

    GPU Support

    1. Verified with nvidia-smi: Tesla T4 GPU available
    2. PyTorch GPU acceleration is enabled"

    Support

    AWS infrastructure support

    AWS Support is a one-on-one, fast-response support channel that is staffed 24x7x365 with experienced and technical support engineers. The service helps customers of all sizes and technical abilities to successfully utilize the products and features provided by Amazon Web Services.

    Similar products

    Customer reviews

    Ratings and reviews

     Info
    0 ratings
    5 star
    4 star
    3 star
    2 star
    1 star
    0%
    0%
    0%
    0%
    0%
    0 reviews
    No customer reviews yet
    Be the first to review this product . We've partnered with PeerSpot to gather customer feedback. You can share your experience by writing or recording a review, or scheduling a call with a PeerSpot analyst.