My colleague Steven Jones wrote the guest post below in order to tell you about an impressive new way to use SAP HANA for large-scale workloads.— Jeff;
Back in May we announced the availability of our new X1 instance type x1.32xlarge, our latest addition to the Amazon EC2 memory-optimized instance family with 2 TB of RAM, purpose built for running large-scale, in-memory applications and in-memory databases like SAP HANA in the AWS cloud.
At the same time, we announced SAP certification for single-node deployments of SAP HANA on X1 and since then many AWS customers have been making use of X1 across the globe for a broad range of HANA OLTP use cases including S/4HANA, Suite on HANA, Business Warehouse on HANA, and other OLAP based BI strategies. Even so, many customers have been asking for the ability to use SAP HANA with X1 instances clustered together in scale-out fashion.
After extensive testing and benchmarking of scale-out HANA clusters in accordance with SAP’s certification processes we’re pleased to announce that today in conjunction with the announcement of BW/4HANA, SAP’s highly optimized next generation business warehouse, our AWS X1 instances are now certified by SAP for large scale-out OLAP deployments including BW/4HANA for up to 7 nodes or 14 TB of RAM. We are excited to be able to support the launch of SAP’s new flagship Business Warehouse offering BW4/HANA with new flexible, scalable, and cost effective deployment options.
Here’s a screenshot from HANA Studio showing a large (14 TB) scale-out cluster running on seven X1 instances:
And this is just the beginning; as indicated, we have plans to make X1 instances available in other sizes and we are testing even larger clusters in the range of 50 TB in our lab. If you need scale-out clusters larger than 14 TB, please contact us; we’d like to work with you.
Reduced Cost and Complexity
Many AWS customers have also been running SAP HANA in scale-out fashion across multiple R3 instances. This new certification brings the ability to consolidate larger scale-out deployments onto fewer larger instances, reducing both cost and complexity. See our SAP HANA Migration guide for details on consolidation strategies.
Flexible High-Availability Options
The AWS platform brings a wide variety of options depending on your needs for ensuring critical SAP HANA deployments like S/4HANA and BW/4HANA are highly available. In fact, customers who have run scale-out deployments of SAP HANA on premises, or with traditional hosting providers, tell us they often have to pay expensive maintenance contracts in addition to purchasing standby nodes or spare hardware to be able to rapidly respond to hardware failures. Others unfortunately forgo this extra hardware and hope nothing happens.
One particularly useful option customers are leveraging on AWS platform is a solution called Amazon EC2 Auto Recovery. Customers simply create an Amazon CloudWatch alarm that monitors their EC2 instance(s) which automatically recovers the instance to a healthy host if it becomes impaired due to an underlying hardware failure or a problem that requires AWS involvement to repair. A recovered instance is identical to the original instance, including attached EBS storage volumes as well as other configurations such as hostname, IP address, and AWS instance IDs. Standard pricing for Amazon CloudWatch applies (for example $0.10 per alarm per month us-east). Essentially this allows you to leverage our spare capacity for rapid recovery while we take care of the unhealthy hardware.
You can deploy your own production ready single-node HANA or scale-out HANA solution on X1 using the updated AWS Quick Start Reference Deployment for SAP HANA in less than an hour using well-tested configurations.
Be sure to also review our SAP HANA Implementation and Operations Guide for other guidance and best practices when planning your SAP HANA implementation on Amazon Web Services.
Are you in the Bay Area on September 7 and want to join us for an exciting AWS and SAP announcement? Register here and we’ll see you in San Francisco!
Can’t make it? Join our livestream on September 7 at 9 AM PST and learn how AWS and SAP are working together to provide value for SAP customers.
We look forward to serving you.
— Steven Jones, Senior Manager, AWS Solutions Architecture
Back with her second guest post, Tina Barr talks about four more hot startups!— Jeff;
This month we are featuring four hot AWS-powered startups:
- Craftsvilla – Offering a platform to purchase ethnic goods.
- SendBird – Helping developers build 1-on-1 messaging and group chat quickly.
- Teletext.io – A solution for content management, without the system.
- Wavefront – A cloud-based analytics platform.
Craftsvilla was born in 2011 out of sheer love and appreciation for the crafts, arts, and culture of India. On a road trip through the Gujarat region of western India, Monica and Manoj Gupta were mesmerized by the beautiful creations crafted by local artisans. However, they were equally dismayed that these artisans were struggling to make ends meet. Monica and Manoj set out to create a platform where these highly skilled workers could connect directly with their consumers and reach a much broader audience. The demand for authentic ethnic products is huge across the globe, but consumers are often unable to find the right place to buy them. Craftsvilla helps to solve this issue.
The culture of India is so rich and diverse, that no one had attempted to capture it on a single platform. Using technological innovations, Craftsvilla combines apparel, accessories, health and beauty products, food items and home décor all in one easily accessible space. For instance, they not only offer a variety of clothing (Salwar suits, sarees, lehengas, and casual wear) but each of those categories are further broken down into subcategories. Consumers can find anything that fits their needs – they can filter products by fabric, style, occasion, and even by the type of work (embroidered, beads, crystal work, handcrafted, etc.). If you are interested in trying new cuisine, Craftsvilla can help. They offer hundreds of interesting products from masalas to traditional sweets to delicious tea blends. They even give you the option to filter through India’s many diverse regions to discover new foods.
Becoming a seller on Craftsvilla is simple. Shop owners just need to create a free account and they’re able to start selling their unique products and services. Craftsvilla’s ultimate vision is to become the ‘one-stop destination’ for all things ethnic. They look to be well on their way!
AWS itself is an engineer on Craftsvilla’s team. Customer experience is highly important to the people behind the company, and an integral aspect of their business is to attain scalability with efficiency. They automate their infrastructure at a large scale, which wouldn’t be possible at the current pace without AWS. Currently, they utilize over 20 AWS services – Amazon Elastic Compute Cloud (EC2), Elastic Load Balancing, Amazon Kinesis, AWS Lambda, Amazon Relational Database Service (RDS), Amazon Redshift, and Amazon Virtual Private Cloud to name a few. Their app QA process will move to AWS Device Farm, completely automated in the cloud, on 250+ services thanks to Lambda. Craftsvilla relies completely on AWS for all of their infrastructure needs, from web serving to analytics.
Check out Craftsvilla’s blog for more information!
After successfully exiting their first startup, SendBird founders John S. Kim, Brandon Jeon, Harry Kim, and Forest Lee saw a great market opportunity for a consumer app developer. Today, over 2,000 global companies such as eBay, Nexon, Beat, Malang Studio, and SK Telecom are using SendBird to implement chat and messaging capabilities on their mobiles apps and websites. A few ways companies are using SendBird:
- 1-on-1 messaging for private messaging and conversational commerce.
- Group chat for friends and interest groups.
- Massive scale chat rooms for live-video streams and game communities.
After graduating from accelerator Y Combinator W16 Batch, the company grew from 1,000,000 monthly chat users to 5,000,000 monthly chat users within months while handling millions of new messages daily across live-video streaming, games, ecommerce, and consumer apps. Customers found value in having a cross-platform, full-featured, and whole-stack approach to a real-time chat API and SDK which can be deployed in a short period of time.
SendBird chose AWS to build a robust and scalable infrastructure to handle a massive concurrent user base scattered across the globe. It uses EC2 with Elastic Load Balancing and Auto Scaling, Route 53, S3, ElastiCache, Amazon Aurora, CloudFront, CloudWatch, and SNS. The company expects to continue partnering with AWS to scale efficiently and reliably.
Marcel Panse and Sander Nagtegaal, co-founders of Teletext.io, had worked together at several startups and experienced the same problem at each one: within the scope of custom software development, content management is a big pain. Even the smallest correction, such as a typo, typically requires a developer, which can become very expensive over time. Unable to find a proper solution that was readily available, Marcel and Sander decided to create their own service to finally solve the issue. Leveraging only the API Gateway, Lambda functions, Amazon DynamoDB, S3, and CloudFront, they built a drop-in content management service (CMS). Their serverless approach for a CMS alternative quickly attracted other companies, and despite intending to use it only for their own needs, the pair decided to professionally market their idea and Teletext.io was born.
Today, Teletext.io is called a solution for content management, without the system. Content distributors are able to edit text and images through a WYSIWYG editor without the help of a programmer and directly from their own website or user interface. There are just three easy steps to get started:
- Include Teletext.io script.
- Add data attributes.
- Login and start typing.
That’s it! There is no system that needs to be installed or maintained by developers – Teletext.io works directly out of the box. In addition to recurring content updates, the data attribution technique can also be used for localization purposes. Making a website multilingual through a CMS can take days or weeks, but Teletext.io can accomplish this task in mere minutes. The time-saving factor is the main benefit for developers and editors alike.
Teletext.io uses AWS in a variety of ways. Since the company is responsible for the website content of others, they must have an extremely fast and reliable system that keeps website visitors from noticing external content being loaded. In addition, this critical infrastructure service should never go down. Both of these requirements call for a robust architecture with as few moving parts as possible. For these reasons, Teletext.io runs a serverless architecture that really makes it stand out. For loading draft content, storing edits and images, and publishing the result, the Amazon API Gateway gets called, triggering AWS Lambda functions. The Lambda functions store their data in Amazon DynamoDB.
Founded in 2013 and based in Palo Alto, Wavefront is a cloud-based analytics platform that stores time series data at millions of points per second. They are able to detect any divergence from “normal” in hybrid and cloud infrastructures before anomalies ever happen. This is a critical service that companies like Lyft, Okta, Yammer, and Box are using to keep running smoothly. From data scientists to product managers, from startups to Fortune 500 companies, Wavefront offers a powerful query engine and a language designed for everyone.
With a pay-as-you-go model, Wavefront gives customers the flexibility to start with the necessary application size and scale up/down as needed. They also include enterprise-class support as part of their pricing at no extra cost. Take a look at their product demos to learn more about how Wavefront is helping their customers.
The Wavefront Application is hosted entirely on AWS, and runs its single-tenant instances and multi-tenant instances in the virtual private cloud (VPC) clusters within AWS. The application has deep, native integrations with CloudWatch and CloudTrail, which benefits many of its larger customers also using AWS. Wavefront uses AWS to create a “software problem”, to operate, automate and monitor clouds using its own application. Most importantly, AWS allows Wavefront to focus on its core business – to build the best enterprise cloud monitoring system in the world.
Here’s the first community-driven edition of the AWS Week in Review. In response to last week’s blog post (AWS Week in Review – Coming Back With Your Help!), 9 other contributors helped to make this post a reality. That’s a great start; let’s see if we can go for 20 this week.
New & Notable Open Source
- cloudhopper is a framework for deploying your APIs with AWS Lambda.
- metalsmith-cloudfront is a Metalsmith plugin for invalidating Amazon CloudFront distributions.
- aws-lambda-container-monitoring monitors the number of active AWS Lambda containers and reports the results to Amazon CloudWatch.
New SlideShare Presentations
- Amazon CloudFront Office Hours – Using Amazon CloudFront with S3 & ELB
- Deep Dive on Amazon S3
- Data Storage for the Long Haul: Compliance and Archive
- Deep Dive on Amazon Aurora
- Deep Dive on Amazon DynamoDB
- Deep Dive on Amazon S3
- Amazon Aurora for Enterprise Database Applications
- August 29 (Oslo, Norway) – Practical encryption of content data using AWS Key Management Service (AWS User Group Norway)
- August 31 (Seattle, WA) – Building Smart Healthcare Applications on AWS
- September 6 (Dublin, Ireland) – Developing with Amazon Alexa – Building Voice Experiences for Services & Devices.
Stay tuned for next week, and please consider helping to make this a community-driven effort!
Amazon CloudWatch helps you to see, diagnose, react to, and resolve issues that arise in your AWS infrastructure and in the applications that you run on AWS. Today, I would like to talk about several usability and functionality improvements to CloudWatch Logs (Store and Monitor OS & Application Log Files with Amazon CloudWatch) and to CloudWatch Dashboards (CloudWatch Dashboards – Create & Use Customized Metrics Views).
Usability Improvements to CloudWatch Logs
CloudWatch Logs is a highly available, scalable, durable, and secure service to manage your operating system and application log files. It allows you to ingest, store, filter, search, and archive the logs, reducing your operational burden and allowing you to focus on your application and your business.
In order to help you to stay efficient and productive even as the number and size of your logs grows, we have made several usability improvements to the CloudWatch Logs Console:
- Improved formatting for log data.
- Simplified access to lengthy log files.
- Easier searching within a log group.
- Simplified collaboration around log files.
- Better searching within a specific time frame.
Prior to today’s launch we also made some improvements to the CloudWatch Dashboards:
- Full screen mode.
- Dark theme.
- Control over range of the Y axis on charts.
- Simplified renaming of charts.
- Persistent storage of chart settings.
CloudWatch Logs Console in Action
Let’s take a look at each of these improvements!
Open up the CloudWatch Logs Console, click on a Log Group, and then on a Log Stream within the group. Find the View options menu on the right:
Click on Expand all in order to see the log messages in expanded, multi-line form like this:
You can also Switch to text view in order to see the logs in their unadorned, plain-text form:
We have also improved the display of log data across all streams within a log group. Once you select a Log Group and click Search Events you can see the log data from all streams with that log group. For example, I can easily identify the Billed Duration for multiple invocations of a single Lambda function:
Even better, we have replaced the original paginated view with an infinite scroll bar. You can now scroll to your heart’s content through log files of any length:
You can now refine your search to a specific time frame or to a custom date range with a single click, like this:
If you are working as part of a team, you can now share the URL of your log analysis session. The URL includes the search parameters and filters, and includes a fragment that looks like this:
These improvements to the CloudWatch Logs Console are available now and you can start using them today. To learn more, read Getting Started with CloudWatch Logs.
Recent Improvements to the CloudWatch Dashboards
You may have already noticed the improvements that we recently made to the CloudWatch Dashboards. First, there’s a new full screen mode for Dashboards, accessible by clicking on Enter full screen in the Actions menu:
Once you are in full screen mode, you can click on Dark to switch to the new, night-owl-friendly dark theme:
Here’s a simple Redis dashboard in full screen mode using the dark theme:
Sometimes you want to have more control over how a chart is displayed on your dashboard. As an example, outliers in your data may make your chart less readable, and you may want to keep the dashboard focused on a specific Y axis range. Here’s a chart where that’s the case; the outlier masks the trend that happened after the big spike:
To edit the Y axis, click on the tool selector and select Edit:
Choose Graph Options and then edit the values for the Y axis until you are satisfied with the appearance of the chart, then click on Update widget:
Here’s what the chart looks like after that:
Many of our customers wanted to be able to rename a chart without leaving the dashboard. You can now do that with a click (hover your mouse near the name and then click on the pencil):
Finally, CloudWatch now remembers the time range, timezone preference, refresh interval, and auto-refresh setting for each chart!
Amazon CloudWatch Partner Ecosystem
I’d like to wrap things up by sharing some of the great work that our partners are doing. The following partners are building value-added solutions on top of CloudWatch:
- Datadog provides integrations to key items in your infrastructure, and gives you the ability to collaborate with your team directly when dealing with incidents.
- Librato provides integrations across elements of your infrastructure, and supports composite metrics and mathematical transformations to time series data.
- SignalFx helps provide you with instant visibility into your metrics, and focuses on data analytics and on delivering alerts on service-wide patterns.
- Splunk offers a platform for operational intelligence that enables you to collect machine data and find insights.
- Sumo Logic is a machine data analytics service for log management and time series metrics that helps you build, run and secure your applications.
If you are a partner and offer something that belongs on this list, let me know and I’ll update it ASAP!— Jeff;
Many AWS customers have been asking us for a way to programmatically analyze their Cost and Usage Reports (read New – AWS Cost and Usage Reports for Comprehensive and Customizable Reporting for more info). These customers are often using AWS to run multiple lines of business, making use of a wide variety of services, often spread out across multiple regions. Because we provide very detailed billing and cost information, this is a Big Data problem and one that can be easily addressed using AWS services!
While I was on vacation earlier this month, we launched a new feature that allows you to upload your Cost and Usage reports to Amazon Redshift and Amazon QuickSight. Now that I am caught up, I’d like to tell you about this feature.
Upload to Redshift
I started by creating a new Redshift cluster (if you already have a running cluster, you need not create another one). Here’s my cluster:
Next, I verified that I had enabled the Billing Reports feature:
Then I hopped over to the Cost and Billing Reports and clicked on Create report:
Next, I named my report (MyReportRedshift), made it Hourly, and enabled support for both Redshift and QuickSight:
I wrapped things up by selecting my delivery options:
I confirmed my desire to create a report on the next page, and then clicked on Review and Complete. The report was created and I was informed that the first report would arrive in the bucket within 24 hours:
While I was waiting I installed PostgreSQL on my EC2 instance (
sudo yum install postgresql94) and verified that I was signed up for the Amazon QuickSight preview. Also, following the directions in Create an IAM Role, I made a read-only IAM role and captured its ARN:
Back in the Redshift console, I clicked on Manage IAM Roles and associated the ARN with my Redshift cluster:
The next day, I verified that the files were arriving in my bucket as expected, and then returned to the console in order to retrieve a helper file so that I could access Redshift:
I clicked on Redshift file and then copied the SQL command:
I inserted the ARN and the S3 region name into the SQL (I had to add quotes around the region name in order to make the query work as expected):
And then I connected to Redshift using
psql (I can use any visual or CLI-based SQL client):
$ psql -h jbcluster.XYZ.us-east-1.redshift.amazonaws.com \ -U root -p 5439 -d dev
Then I ran the SQL command. It created a pair of tables and imported the billing data from S3.
Querying Data in Redshift
Using some queries supplied by my colleagues as a starting point, I summed up my S3 usage for the month:
And then I looked at my costs on a per-AZ basis:
And on a per-AZ, per-service basis:
Just for fun, I spent some time examining the Redshift Console. I was able to see all of my queries:
Analyzing Data with QuickSight
I also spent some time analyzing the cost and billing data using Amazon QuickSight. I signed in and clicked on Connect to another data source or upload a file:
Then I dug in to my S3 bucket (
jbarr-bcm) and captured the URL of the manifest file (
I selected S3 as my data source and entered the URL:
QuickSight imported the data in a few seconds and the new data source was available. I loaded it into SPICE (QuickSight’s in-memory calculation engine). With three or four more clicks I focused on the per-AZ data, and excluded the data that was not specific to an AZ:
Another click and I switched to a pie chart view:
I also examined the costs on a per-service basis:
As you can see, the new data and the analytical capabilities of QuickSight allow me (and you) to dive deep into your AWS costs in minutes.
This new feature is available now and you can start using it today!
Back in 2012 I realized that something interesting happened in AWS-land just about every day. In contrast to the periodic bursts of activity that were the norm back in the days of shrink-wrapped software, the cloud became a place where steady, continuous development took place.
In order to share all of this activity with my readers and to better illustrate the pace of innovation, I published the first AWS Week in Review in the spring of 2012. The original post took all of about 5 minutes to assemble, post and format. I got some great feedback on it and I continued to produce a steady stream of new posts every week for over 4 years. Over the years I added more and more content generated within AWS and from the ever-growing community of fans, developers, and partners.
Unfortunately, finding, saving, and filtering links, and then generating these posts grew to take a substantial amount of time. I reluctantly stopped writing new posts early this year after spending about 4 hours on the post for the week of April 25th.
After receiving dozens of emails and tweets asking about the posts, I gave some thought to a new model that would be open and more scalable.
The AWS Week in Review is now a GitHub project (https://github.com/aws/aws-week-in-review). I am inviting contributors (AWS fans, users, bloggers, and partners) to contribute.
Every Monday morning I will review and accept pull requests for the previous week, aiming to publish the Week in Review by 10 AM PT. In order to keep the posts focused and highly valuable, I will approve pull requests only if they meet our guidelines for style and content.
At that time I will also create a file for the week to come, so that you can populate it as you discover new and relevant content.
Content & Style Guidelines
Here are the guidelines for making contributions:
- Relevance -All contributions must be directly related to AWS.
- Ownership – All contributions remain the property of the contributor.
- Validity – All links must be to publicly available content (links to free, gated content are fine).
- Timeliness – All contributions must refer to content that was created on the associated date.
- Neutrality – This is not the place for editorializing. Just the facts / links.
I generally stay away from generic news about the cloud business, and I post benchmarks only with the approval of my colleagues.
And now a word or two about style:
- Content from this blog is generally prefixed with “I wrote about POST_TITLE” or “We announced that TOPIC.”
- Content from other AWS blogs is styled as “The BLOG_NAME wrote about POST_TITLE.”
- Content from individuals is styled as “PERSON wrote about POST_TITLE.”
- Content from partners and ISVs is styled as “The BLOG_NAME wrote about POST_TITLE.”
There’s room for some innovation and variation to keep things interesting, but keep it clean and concise. Please feel free to review some of my older posts to get a sense for what works.
Over time we might want to create a more compelling visual design for the posts. Your ideas (and contributions) are welcome.
Over the years I created the following sections:
- Daily Summaries – content from this blog, other AWS blogs, and everywhere else.
- New & Notable Open Source.
- New SlideShare Presentations.
- New YouTube Videos including APN Success Stories.
- New AWS Marketplace products.
- New Customer Success Stories.
- Upcoming Events.
- Help Wanted.
Some of this content comes to my attention via RSS feeds. I will post the OPML file that I use in the GitHub repo and you can use it as a starting point. The New & Notable Open Source section is derived from a GitHub search for aws. I scroll through the results and pick the 10 or 15 items that catch my eye. I also watch /r/aws and Hacker News for interesting and relevant links and discussions.
Over time, it is possible that groups or individuals may become the primary contributor for a section. That’s fine, and I would be thrilled to see this happen. I am also open to the addition to new sections, as long as they are highly relevant to AWS.
Adding Content / Creating a Pull Request
It is very easy to participate in this process. You don’t need to use any shell commands or text editors. Start by creating a GitHub account and logging in. I set up two-factor authentication for my account and you might want to do the same.
Now, find a piece of relevant content. As an example, I’ll use the presentation Amazon Aurora for Enterprise Database Applications. I visit the current
aws-week-in-review file and click on the Edit button (the pencil icon):
Then I insert the new content (line 81):
I could have inserted several pieces of new content if desired.
Next, I enter a simple commit message, indicate that the commit should go to a branch (this is important), and click on Propose file change.
And that’s it! In my role as owner of the file, I’ll see the pull request, review it, and then merge it in to the master branch.
Earlier this year I tried to automate the process, but I did not like the results. You are welcome to give this a shot on your own. I do want to make sure that we continue to exercise human judgement in order to keep the posts as valuable as possible.
Let’s Do It
I am super excited about this project and I cannot wait to see those pull requests coming in. Please let me know (via a blog comment) if you have any suggestions or concerns.
I should note up front that I am very new to Git-based collaboration and that this is going to be a learning exercise for me. Do not hesitate to let me know if there’s a better way to do things!— Jeff;
In my recent post, I Love My Amazon WorkSpace, I shared the story of how I became a full-time user and big fan of Amazon WorkSpaces. Since writing the post I have heard similar sentiments from several other AWS customers.
Today I would like to tell you about some new and recent developments that will make WorkSpaces more economical, more flexible, and more useful:
- Hourly WorkSpaces – You can now pay for your WorkSpace by the hour.
- Expanded Root Volume – Newly launched WorkSpaces now have an 80 GB root volume.
Let’s take a closer look at these new features.
If you only need part-time access to your WorkSpace, you (or your organization, to be more precise) will benefit from this feature. In addition to the existing monthly billing, you can now use and pay for a WorkSpace on an hourly basis, allowing you to save money on your AWS bill. If you are a part-time employee, a road warrior, share your job with another part-timer, or work on multiple short-term projects, this feature is for you. It is also a great fit for corporate training, education, and remote administration.
There are now two running modes – AlwaysOn and AutoStop:
- AlwaysOn – This is the existing mode. You have instant access to a WorkSpace that is always running, billed by the month.
- AutoStop – This is new. Your WorkSpace starts running and billing when you log in, and stops automatically when you remain disconnected for a specified period of time.
A WorkSpace that is running in AutoStop mode will automatically stop a predetermined amount of time after you disconnect (1 to 48 hours). Your WorkSpaces Administrator can also force a running WorkSpace to stop. When you next connect, the WorkSpace will resume, with all open documents and running programs intact. Resuming a stopped WorkSpace generally takes less than 90 seconds.
Your WorkSpaces Administrator has the ability to choose your running mode when launching your WorkSpace:
The Administrator can change the AutoStop time and the running mode at any point during the month. They can also track the number of working hours that your WorkSpace accumulates during the month using the new UserConnected CloudWatch metric, and switch from AutoStop to AlwaysOn when this becomes more economical. Switching from hourly to monthly billing takes place upon request; however, switching the other way takes place at the the start of the following month.
All new Amazon WorkSpaces can take advantage of hourly billing today. If you’re using a custom image for your WorkSpaces, you’ll need to refresh your custom images from the latest Amazon WorkSpaces bundles. The ability for existing WorkSpaces to switch to hourly billing will be added in the future.
To learn more about pricing for hourly WorkSpaces, visit the WorkSpaces Pricing page.
Expanded Root Volume
By popular demand we have expanded the size of the root volume for newly launched WorkSpaces to 80 GB, allowing you to run more applications and store more data at no additional cost. Your WorkSpaces Administrator can rebuild existing WorkSpaces in order to upgrade them to the larger root volumes (read Rebuild a WorkSpace to learn more). Rebuilding a WorkSpace will restore the root volume (C:) to the most recent image of the bundle that was used to create the WorkSpace. It will also restore the data volume (D:) from the last automatic snapshot.
Some WorkSpaces Resources
While I have your attention, I would like to let you know about a couple of other important WorkSpaces resources:
- Getting Started – Our WorkSpaces Getting Started page includes a new step-by-step WorkSpaces Implementation Guide and other handy documentation.
- Recorded Webinar – Late last month my colleague Salman Paracha delivered the Intro to Amazon WorkSpaces webinar. Watch it to learn how WorkSpaces can be used to support a diverse and dynamic global workforce and improve your organization’s security position while providing users with a familiar and productive desktop experience.
- Whitepapers – Our new Best Practices for Deploying Amazon WorkSpaces whitepaper addresses network considerations, directory services for user authentication, security, monitoring, and logging. The Desktop-as-a-Service whitepaper reviews the greater operational control, reduced costs, and security benefits of WorkSpaces.
- Case Studies – The Louisiana Department of Corrections and Endemol Shine Group case studies will show you how AWS customers are putting WorkSpaces to use.
The features that I described above are available now and you can start using them today!
Everyone on the AWS team understands the value of educating our customers on the best ways to use our services. We work hard to create documentation, training materials, and blog posts for you! We run live events such as our Global AWS Summits and AWS re:Invent where the focus is on education. Last but not least, we put our heads together and create a fresh lineup of webinars for you each and every month.
We have a great selection of webinars on the schedule for August. As always they are free, but they do fill up and I strongly suggest that you register ahead of time. All times are PT, and each webinar runs for one hour:
- 9:00 AM – Introducing Amazon EMR Release 5.0: Faster, Easier, Hadoop, Spark, and Presto.
- 10:30 AM – Best Practices for Data Center Migration Planning.
- Noon – Best Practices for Running SAP HANA Workloads with Amazon EC2 X1 Instances.
- 9:00 AM – Amazon Aurora for the Enterprise: Lower Cost, Better Performance.
- 10:30 AM – Getting Started with Serverless Architectures.
- Noon – Best Practices for Building a Data Lake with Amazon S3.
- 9:00 AM – Managing IoT and Time Series Data with Amazon ElastiCache for Redis.
- 10:30 AM – Running Microservices and Docker on AWS Elastic Beanstalk.
- Noon – Getting Started with Microsoft SQL Server 2016 on Amazon EC2.
- 9:00 AM – Introduction to Amazon Kinesis Firehose.
- 10:30 AM – Building Serverless Chat Bots.
- Noon – Continuous Delivery to Amazon ECS.
- 9:00 AM – Stream Data Analytics with Amazon Kinesis Firehose & Redshift.
- 10:30 AM – Getting Started with AWS Device Farm.
- Noon – AWS IoT Button.
PS – Check out the AWS Webinar Archive for more great content!
Today I would like to tell you about a new AWS Solution. This one is cool because of what it does and how it works! Like the AWS Quick Starts, this one was built by AWS Solutions Architects and incorporates best practices for security and high availability.
The new Transit VPC Solution shows you how to implement a very useful networking construct that we call a transit VPC. You can use this to connect multiple Virtual Private Clouds (VPCs) that might be geographically disparate and/or running in separate AWS accounts, to a common VPC that serves as a global network transit center. This network topology simplifies network management and minimizes the number of connections that you need to set up and manage. Even better, it is implemented virtually and does not require any physical network gear or a physical presence in a colocation transit hub. Here’s what this looks like:
In this diagram, the transit VPC is central, surrounded by additional “spoke” VPCs, corporate data centers, and other networks.
The transit VPC supports several important use cases:
- Private Networking – You can build a private network that spans two or more AWS Regions.
- Shared Connectivity – Multiple VPCs can share connections to data centers, partner networks, and other clouds.
- Cross-Account AWS Usage – The VPCs and the AWS resources within them can reside in multiple AWS accounts.
The solution uses a AWS CloudFormation stack to launch and configure all of the AWS resources. It provides you with three throughput options ranging from 500 Mbps to 2 Gbps, each implemented over a pair of connections for high availability. The stack makes use of the Cisco Cloud Services Router (CSR), which is now available in AWS Marketplace. You can use your existing CSR licenses (the BYOL model) or you can pay for your CSR usage on an hourly basis. The cost to run a transit VPC is based on the throughput option and licensing model that you choose, and ranges from $0.21 to $8.40 per hour, with an additional cost (for AWS resources) of $0.10 per hour for each spoke VPC. There’s an additional cost of $1 per month for a AWS Key Management Service (KMS) customer master key that is specific to the solution. All of these prices are exclusive of network transit costs.
The template installs and uses a pair of AWS Lambda functions in a creative way!
The VGW Poller function runs every minute. It scans all of the AWS Regions in the account, looking for appropriately tagged Virtual Private Gateways in spoke VPCs that do not have a VPN connection. When it finds one, it creates (if necessary) the corresponding customer gateway and the VPN connections to the CSR, and then saves the information in an S3 bucket.
The Cisco Configurator function is triggered by the Put event on the bucket. It parses the VPN connection information and generates the necessary config files, then pushes them to the CSR instances using SSH. This allows the VPN tunnels to come up and (via the magic of BGP), neighbor relationships will be established with the spoke VPCs.
By using Lambda in this way, new spoke VPCs can be brought online quickly without the overhead of keeping an underutilized EC2 instance up and running.
The solution’s implementation guide, as always, contains step-by-step directions and security recommendations.— Jeff;
PS – Check out additional network best practice guidance to find answers to common network questions!
We introduced AWS Import/Export Snowball last fall from the re:Invent stage. The Snowball appliance is designed for customers who need to transfer large amounts of data into or out of AWS on a one-time or recurring basis (read AWS Import/Export Snowball – Transfer 1 Petabyte Per Week Using Amazon-Owned Storage Appliances to learn more).
Today we are launching two important additions to Snowball. Here’s the scoop:
- Snowball Job Management API – The new Snowball API lets you build applications that create and manage Snowball jobs.
- S3 Adapter – The new Snowball S3 Adapter lets you access a Snowball appliance as if it were an S3 endpoint.
Time to dive in!
Snowball Job Management API
The original Snowball model was interactive and console-driven. You could create a job (basically “Send me a Snowball”) and then monitor its progress, tracking the shipment, transit, delivery, and return to AWS visually. This was great for one-off jobs, but did not meet the needs of customers who wanted to integrate Snowball into their existing backup or data transfer model. Based on the requests that we received from these customers and from our Storage Partners, we are introducing a Snowball Job Management API today.
The Snowball Job Management API gives our customers and partners the power to make Snowball an intrinsic, integrated part of their data management solutions. Here are the primary functions:
CreateJob– Create an import or export job & initiates shipment of an appliance.
ListJobs– Fetch a list of jobs and associated job states.
DescribeJob– Fetch information about a specific job.
Read the API Reference to learn more!
I’m looking forward to reading about creative and innovative applications that make use of this new API! Leave me a comment and let me know what you come up with.
The new Snowball S3 Adapter allows you to access a Snowball as if it were an Amazon S3 endpoint running on-premises. This allows you to use your existing, S3-centric tools to move data to or from a Snowball.
The adapter is available for multiple Linux distributions and Windows releases, and is easy to install:
- Download the appropriate file from the Snowball Tools page and extract its contents to a local directory.
- Verify that the adapter’s configuration is appropriate for your environment (the adapter listens on port 8080 by default).
- Connect your Snowball to your network and get its IP address from the built-in display on the appliance.
- Visit the Snowball Console to obtain the unlock code and the job manifest.
- Launch the adapter, providing it with the IP address, unlock code, and manifest file.
With the adapter up and running, you can use your existing S3-centric tools by simply configuring them to use the local endpoint (the IP address of the on-premises host and the listener port). For example, here’s how you would run the
s3 ls command on the on-premises host:
$ aws s3 ls --endpoint http://localhost:8080
After you copy your files to the Snowball, you can easily verify that the expected number of files were copied:
$ snowball validate
The initial release of the adapter supports a subset of the S3 API including GET on buckets and on the service, HEAD on a bucket and on objects, PUT and DELETE on objects, and all of the multipart upload operations. If you plan to access the adapter using your own code or third party tools, some testing is advisable.
To learn more, read about the Snowball Transfer Adapter.
These new features are available now and you can start using them today!