AWS Official Blog
A few months ago I announced that you would soon be able to access resources in a VPC from your AWS Lambda functions. I am happy to announce that this much-wanted feature is now available and that you can start using it today!
Your Lambda functions can now access Amazon Redshift data warehouses, Amazon ElastiCache clusters, Amazon Relational Database Service (RDS) instances, and service endpoints that are accessible only from within a particular VPC. In order to do this, you simply select one of your VPCs and identify the relevant subnets and security groups. Lambda uses this information to set up elastic network interfaces (ENIs) and private IP addresses (drawn from the subnet or subnets that you specified) so that your Lambda function has access to resources in the VPC.
Accessing Resources in a VPC
You can set this up when you create a new function. You can also update an existing function so that it has VPC access. You can configure this feature from the Lambda Console or from the CLI. Here’s how you set it up from the Console:
That’s all you need to do! Be sure to read Configuring a Lambda Function to Access Resources in an Amazon VPC in the Lambda documentation if you have any questions.
Things to Know
Here are a couple of things that you should know about this new feature:
ENI & IP Address Resources – Because Lambda automatically scales based on the number of events that is needs to process, your VPC must have an adequate supply of free IP addresses on the designated subnets.
Internet Access – As soon as you enable this functionality for a particular function, the function no longer has access to the Internet by default. If your function requires this type of access, you will need to set up a Managed NAT Gateway in your VPC (see New – Managed NAT (Network Address Translation) Gateway for AWS for more information) or run your own NAT (see NAT Instances).
Security Groups – The security groups that you choose for a function will control the function’s access to the resources in the subnets and on the Internet.
S3 Endpoints – You can also use this feature to access S3 endpoints within a VPC (consult New – VPC Endpoint for Amazon S3 to learn more).
Webinar – To learn more about this new feature, join our upcoming webinar, Essentials: Introducing AWS VPC Support for AWS Lambda.— Jeff;
We want to make it as easy as possible for you to secure your AWS environment. Some of our more recent announcements in this area include encrypted EBS boot volumes, encryption at rest for Amazon Aurora, and support for AWS Key Management Service (KMS) across several different services.
Today we are giving you some additional options for data stored in Amazon Relational Database Service (RDS). You can now share encrypted database snapshots with other AWS accounts. You can also add encryption to a previously unencrypted database instance.
Sharing Encrypted Snapshots
When you are using encryption at rest for a database instance, automatic and manual database snapshots of the instance are also encrypted. Up until now, encrypted snapshots were private to a single AWS account and could not be shared. Today we are giving you the ability to share encrypted snapshots with up to 20 other AWS accounts. You can do this from the AWS Management Console, AWS Command Line Interface (CLI), or via the RDS API. You can share encrypted snapshots within an AWS region, but you cannot share them publicly. As is the case with the existing sharing feature, today’s release applies to manual snapshots.
To share an encrypted snapshot, select it and click on Share Snapshot. This will open up the Manage Snapshot Permissions page. Enter one or more account IDs (click on Add after each one) and click on Save when you have entered them all:
The accounts could be owned by your organization (perhaps you have separate accounts for dev, test, staging, and production) or by your business partners. Backing up to your mission-critical databases to a separate AWS account is a best practice, and one that you can implement using this new feature while also gaining the benefit of encryption at rest.
After you click on Save, the other accounts have access to the shared snapshots. The easiest way to locate them is to visit the RDS Console and filter the list using Shared with Me:
The snapshot can be used to create a new RDS database instance. To learn more, read about Sharing a Database Snapshot.
Adding Encryption to Existing Database Instances
You can now add encryption at rest using KMS keys to a previously unencrypted database instance. This is a simple, multi-step process:
- Create a snapshot of the unencrypted database instance.
- Copy the snapshot to a new, encrypted snapshot. Enable encryption and specify the desired KMS key as you do so:
- Restore the encrypted snapshot to a new database instance:
- Update your application to refer to the endpoint of the new database instance:
And that’s all you need to do! You can use a similar procedure to change encryption keys for existing database instances. To learn more, read about Copying a Database Snapshot.— Jeff;
Earlier this year I told you about the AWS IoT Mega Contest. The contest closed at the end of January, the judges retired to our secret lair deep in the heart of Seattle, and we have chosen the winners. There were an impressive number of equally impressive projects and it was not easy to evaluate them against our criteria and to pick our favorites. After extended deliberation, we managed to choose ten projects. With no further fanfare, here we go!
The following two entrants will receive the first prize, a Kindle Fire HD 10:
- Chris Synan – Voice Controlled Drone with Amazon Echo and 3DR IRIS+.
- Marian Mihailescu – Baby NAP (Night Activity Program).
The following three entrants will receive the second prize, an Amazon Echo:
- Arlo Carreon – Magic Mirror – Amazon Echo and AWS IoT.
- Kittitouch Areerob – Automatic Vehicle Accident Alert System using AWS IoT.
- Tan See Youu – Crying AED.
The following five entrants will receive the third prize, Amazon Fire TV Gaming Edition:
- Daniel Porrey – AWS IoT Button for LIFX.
- Demirhan Aydin – WaterPi: Houseplant Remote Watering and Monitoring System.
- truthkos – Ambient Light Sensor for Philips Hue.
- Aaron Parecki – Magic Checkin Clock.
- Carmelito Andrade – Plant Monitoring System using AWS IoT.
What I Learned
After spending time examining the entries in detail, I came away impressed by a couple of things. To wit:
- There are lots of creative people out there! The initial ideas and the resulting projects were literally all over the map.
- IoT is here now. People are building devices, sites, and applications that are sophisticated and useful.
- Connecting to and working within the real world is a lot harder than running within the clean, abstract confines of a virtual machine. A successful IoT application must be prepared to deal with erroneous or missing data, intermittent connections, and more.
- Building these applications requires and exercises a diverse set of skills. In addition to creativity, a successful IoT project can require theoretical & practical electronics skills (both analog and digital), 3D modeling & printing, along with the ability to write code that runs on small devices and in the cloud, generally using multiple languages, frameworks, and cloud services! If you currently have a subset of these skills, jumping in to IoT is a great way to put them to use.
Building world-class games is a very difficult, time-consuming, and expensive process. The audience is incredibly demanding. They want engaging, social play that spans a wide variety of desktop, console, and mobile platforms. Due to the long lead time inherent in the game development and distribution process, the success or failure of the game can often be determined on launch day, when pent-up demand causes hundreds of thousands or even millions of players to sign in and take the game for a spin.
Behind the scenes, the development process must be up to this challenge. Game creators must be part of a team that includes developers with skills in story telling, game design, physics, logic design, sound creation, graphics, visual effects, and animation. If the game is network-based, the team must also include expertise in scaling, online storage, network communication & management, security.
With development and creative work that can take 18 to 36 months, today’s games represent a considerable financial and reputational risk for the studio. Each new game is a make-or-break affair.
New AWS Game Services
Today I would like to tell you about a pair of new AWS products that are designed for use by professional game developers building cloud-connected, cross-platform games. We started with several proven, industry leading engines and developer tools, added a considerable amount of our own code, and integrated the entire package with our Twitch video platform and community, while also mixing in access to relevant AWS messaging, identity, and storage services. Here’s what we are announcing today:
Lumberyard – A game engine and development environment designed for professional developers. A blend of new and proven technologies from CryEngine, Double Helix, and AWS, Lumberyard simplifies and streamlines game development. As a game engine, it supports development of cloud-connected and standalone 3D games, with support for asset management, character creation, AI, physics, audio, and more. On the development side, the Lumberyard IDE allows you to design indoor and outdoor environments, starting from a blank canvas. You (I just promoted you to professional game developer) can take advantage of built-in content workflows and an asset pipeline, editing game assets in Photoshop, Maya, or 3ds Max for editing and bringing them in to the IDE afterward. You can program your game in the traditional way using C++ and Visual Studio (including access to the AWS SDK for C++) or you can use our Flow Graph tool and the cool new Cloud Canvas to create cloud-connected gameplay features using visual scripting.
Amazon GameLift – Many modern games include a server or backend component that must scale in proportion to the number of active sessions. Amazon GameLift will help you to deploy and scale session-based multiplayer game servers for the games that you build using Lumberyard. You simply upload your game server image to AWS and deploy the image into a fleet of EC2 instances that scales up as players connect and play. You don’t need to invest in building, scaling, running, or monitoring your own fleet of servers. Instead, you pay a small fee per daily active user (DAU) and the usual EC2 On-Demand rates for the compute capacity, EBS storage, and bandwidth that your users consume.
Twitch Integration – Modern gamers are a very connected bunch. When they are not playing themselves, they like to connect and interact with other players and gaming enthusiasts on Twitch. Professional and amateur players display their talents on Twitch and create large, loyal fan bases. In order to take this trend even further and to foster the establishment of deeper connections and stronger communities, games built with Lumberyard will be able to take advantage of two new Twitch integration features. Twitch ChatPlay allows you to build games that respond to keywords in a Twitch chat stream. For example, the audience can vote to have the player take the most desired course of action. Twitch JoinIn allows a broadcaster to invite a member of the audience into to the game from within the chat channel.
These services, like many other parts of AWS, are designed to allow you to focus on the unique and creative aspects of your game, with an emphasis on rapid turnaround and easy iteration so that you can continue to hone your gameplay until it reaches the desired level of engagement and fun.
Support Services – As the icing on this cake, we are also launching a range of support options including a dedicated Lumberyard forum and a set of tutorials (both text and video). Multiple tiers of paid AWS support are also available.
Developing with Lumberyard
Lumberyard is at the heart of today’s announcement. As I mentioned earlier, it is designed for professional developers and supports development of high-quality, cross-platform games. We are launching with support for the following environments:
- Windows – Vista, Windows 7, 8, and 10.
- Console – PlayStation 4 and Xbox One.
Support for mobile devices and VR headsets is in the works and should be available within a couple of months.
The Lumberyard development environment runs on your Windows PC or laptop. You’ll need a fast, quad-core processor, at least 8 GB of memory, 200 GB of free disk space, and a high-end video card with 2 GB or more of memory and Direct X 11 compatibility. You will also need Visual Studio 2013 Update 4 (or newer) and the Visual C++ Redistributables package for Visual Studio 2013.
The Lumberyard Zip file contains the binaries, templates, assets, and configuration files for the Lumberyard Editor. It also includes binaries and source code for the Lumberyard game engine. You can use the engine as-is, you can dig in to the source code for reference purposes, or you can customize it in order to further differentiate your game. The Zip file also contains the Lumberyard Launcher. This program makes sure that you have properly installed and configured Lumberyard and the third party runtimes, SDKs, tools, and plugins.
The Lumberyard Editor encapsulates the game under development and a suite of tools that you can use to edit the game’s assets.
The Lumberyard Editor includes a suite of editing tools (each of which could be the subject of an entire blog post) including an Asset Browser, a Layer Editor, a LOD Generator, a Texture Browser, a Material Editor, Geppetto (character and animation tools), a Mannequin Editor, Flow Graph (visual programming), an AI Debugger, a Track View Editor, an Audio Controls Editor, a Terrain Editor, a Terrain Texture Layers Editor, a Particle Editor, a Time of Day Editor, a Sun Trajectory Tool, a Composition Editor, a Database View, and a UI Editor. All of the editors (and much more) are accessible from one of the toolbars at the top.
In order to allow you to add functionality to your game in a selective, modular form, Lumberyard uses a code packaging system that we call Gems. You simply enable the desired Gems and they’ll be built and included in your finished game binary automatically. Lumberyard includes Gems for AWS access, Boids (for flocking behavior), clouds, game effects, access to GameLift, lightning, physics, rain, snow, tornadoes, user interfaces, multiplayer functions, and a collection of woodlands assets (for detailed, realistic forests).
Coding with Flow Graph and Cloud Canvas
Traditionally, logic for games was built by dedicated developers, often in C++ and with the usual turnaround time for an edit/compile/run cycle. While this option is still open to you if you use Lumberyard, you also have two other options: Lua and Flow Graph.
Flow Graph is a modern and approachable visual scripting system that allows you to implement complex game logic without writing or or modifying any code. You can use an extensive library of pre-built nodes to set up gameplay, control sounds, and manage effects.
Flow graphs are made from nodes and links; a single level can contain multiple graphs and they can all be active at the same time. Nodes represent game entities or actions. Links connect the output of one node to the input of another one. Inputs have a type (Boolean, Float, Int, String, Vector, and so forth). Output ports can be connected to an input port of any type; an automatic type conversion is performed (if possible).
There are over 30 distinct types of nodes, including a set (known as Cloud Canvas) that provide access to various AWS services. These include two nodes that provide access to Amazon Simple Queue Service (SQS), four nodes that provide access to Amazon Simple Notification Service (SNS), seven nodes that provide read/write access to Amazon DynamoDB, one to invoke an AWS Lambda function, and another to manage player credentials using Amazon Cognito. All of the games calls to AWS are made via an AWS Identity and Access Management (IAM) user that you configure in to Cloud Canvas.
Here’s a node that invokes a Lambda function named DailyGiftLambda:
Here is a flow graph that uses Lambda and DynamoDB to implement a “Daily Gift” function:
As usual, I have barely scratched the surface here! To learn more, read the Cloud Canvas documentation in the Lumberyard User Guide.
Deploying With Amazon GameLift
If your game needs a scalable, cloud-based runtime environment, you should definitely take a look at Amazon GameLift.
You can use it to host many different types of shared, connected, regularly-synchronized games including first-person shooters, survival & sandbox games, racing games, sports games, and MOBA (multiplayer Online Battlefield Arena) games.
After you build your server-side logic, you simply upload it to Amazon GameLift. It will be converted to a Windows-based AMI (Amazon Machine Image) in a matter of minutes. Once the AMI is ready, you can create an Amazon GameLift fleet (or a new version of an existing one), point it at the AMI, and your backend will be ready to go.
Your fleets, and the game sessions, running on each fleet, are visible in the Amazon GameLift Console:
Your Flow Graph code can use the GameLift Gem to create an Amazon GameLift session and to start the session service.
To learn more, consult the Amazon GameLift documentation.
Last but definitely not least, your games can integrate with Twitch via Twitch ChatPlay and Twitch JoinIn.
As I mentioned earlier, you can create games that react to keywords entered in a designated Twitch channel. For example, here’s a Flow Graph that listens for the keywords red, yellow, blue, green, orange, and violet.
Pricing and Availability
Lumberyard and Amazon GameLift are available now and you can start building your games today!
You can build and run connected and standalone games using Lumberyard at no charge. You are responsible for the AWS charges for any calls made to AWS services using the IAM user configured in to Cloud Canvas, or through calls made using the AWS SDK for C++, along with any charges for the use of GameLift.
Amazon GameLift is launching in the US East (Northern Virginia) and US West (Oregon) regions, and will be coming to other AWS regions as well. As part of AWS Free Usage tier, you can run a fleet comprised of one c3.large instance for up to 125 hours per month for a period of one year. After that, you pay the usual On-Demand rates for the EC2 instances that you use, plus the charge for 50 GB / month of EBS storage per instance, and $1.50 per month for every 1000 daily active users.— Jeff;
Let’s take a quick look at what happened in AWS-land last week:
New & Notable Open Source
- goad is an AWS Lambda powered, highly distributed, load testing tool.
- python-lambder lets you create and manage scheduled AWS Lambdas from the command line.
- sevenseconds is an AWS account configurator.
- Zappa implements serverless WSGI with AWS Lambda and API Gateway.
- ctl-transcode transcodes videos using AWS.
- AWS is a set of PowerShell scripts, functions, and modules for managing AWS.
- ssc-lambda is a set of SSC Lambda functions for AWS processing.
- rifactor can automatically refactor your AWS Reserved Instances to match your running instances.
- Eureka is an AWS service registry for resilient mid-tier load balancing and failover.
- lamvery is a function-based deployment and management tool for AWS Lambda.
New SlideShare Presentations
- AWS Security Day:
- AWS January 2016 Webinars:
- Introduction to Docker on AWS.
- Amazon Aurora for Enterprise Database Applications.
- Real-World Smart Applications with Amazon Machine Learning.
- Infrastructure as Code.
- Best Practices for Building IoT Backends with AWS IoT and AWS Lambda.
- Cloud Data Migration: 6 Strategies for Getting Data into AWS.
- AWS IoT – Getting Started.
- Getting Started with Big Data: Analytic Options on AWS & Common Use Cases.
- Introduction to Deploying Applications on AWS.
New Customer Success Stories
- Career Builder – By automating its software release process using AWS CodePipeline and AWS CodeDeploy, CareerBuilder increases update speed and assures quality code, freeing up developers to focus on the core product.
- edotco Group – edotco Group has achieved availability well in excess of its service level agreement of 99.95 percent and reduced infrastructure costs by at least 50 percent over five years using AWS.
- KeptMe – By launching its service on AWS, KeptMe was able to quickly expand to more than 4,000 schools in nine different countries.
- Open Universities Australia – By moving its collocated data center to AWS, OUA reduced the the time required to deliver changes to production from three months to less than two hours, cut costs by up to AU$1 million (US$726,850) over two years, and improved the performance of its websites by up to 20 percent.
- Sokrati – By using AWS, Sokrati reduced the data in its database from 20 terabytes to 2 terabytes and reduced its infrastructure costs by 35 percent.
- Time Inc. – Time Inc. uses AWS Enterprise Support to assist with planning and executing the migration of existing and new applications to AWS.
- 91App – Using AWS has enabled 91App to create and launch digital campaigns in just 24 hours, compared to the several weeks the same processes would have required with a physical IT infrastructure.
- Air Works – By using AWS, Air Works has improved its operational performance by 84 percent and its response times by 160 percent.
- Autodesk – Autodesk can monitor and control the use of hundreds of AWS accounts from a single pane of glass.
- GENALICE – GENALICE uses AWS to run the Population Calling module of its GENALICE MAP Next-Generation Sequencing data analysis suite.
- The Guardian – Guardian News and Media increased the velocity of releases for its digital properties to 40,000 in 2015, up from 25 in 2012, by using AWS.
- Jelly Button Games – Jelly Button Games, an Israeli social gaming company, can grow its business while its AWS environment handles up to one million game server requests a minute.
- Lyft – By using Spot, the startup saves up to 75 percent monthly versus on-demand instances for routine testing processes that do not require the most current or most powerful compute resources.
New YouTube Videos
- Introduction to Amazon Route 53.
- Introduction to AWS Training & Certification.
- Introduction to AWS Snowball.
- Container Day:
- February 11 – Partner Webinar – Finding the Hidden Waste in Your AWS Infrastructure.
- AWS Loft – San Francisco.
- AWS Loft – New York.
- AWS Global Summit Series.
Let’s take a quick look at what happened in AWS-land last week:
My colleague Mike Stroh is part of our sales training team. He wrote the guest post below to introduce you to our newest AWS training courses.— Jeff;
We routinely tweak our 3-day AWS technical training courses to keep pace with AWS platform updates, incorporate learner feedback, and the latest best practices.
Today I want to tell you about some exciting enhancements to Developing on AWS. Whether you’re moving applications to AWS or developing specifically for the cloud, this course can show you how to use the AWS SDK to create secure, scalable cloud applications that tap the full power of the platform.
We’ve made a number of updates to the course—most stem directly from the experiences and suggestions of developers who took previous versions of the course. Here are some highlights of what’s new:
- Balance of Concepts and Code – The updated course expands coverage of key concepts, best practices, and troubleshooting tips for AWS services to help students build a mental model before diving into code. Students then use an AWS SDK to develop apps that apply these concepts in hands-on labs.
- AWS SDK Labs – Practice labs are designed to emphasize the AWS SDK, reflecting how developers actually work and create solutions. Lab environments now include EC2 instances preloaded with all required programming language SDKs, developer tools, and IDEs. Students can simply log in and start learning!
- Relevant to More Developers – The additional programming language support helps make the course more useful to both startup and enterprise developers.
- Expanded Coverage of Developer-Oriented AWS Services – The updated course put more focus on the AWS services relevant to application development. So there’s expanded coverage of Amazon DynamoDB, plus new content on AWS Lambda, Amazon Cognito, Amazon Kinesis Streams, Amazon ElastiCache, AWS CloudFormation, and others.
Here’s a map that will help you to understand how the course flows from topic to topic:
— Mike Stroh, Content & Community Manager
Regular readers of this blog will know that I am a huge fan of Amazon WorkSpaces. In fact, after checking my calendar, I verified that every blog post I have written in the last 10 months has been done from within my WorkSpace. Regardless of my location—office, home, or hotel room—performance, availability, and functionality have all been excellent. Until you have experienced a persistent, cloud-based desktop for yourself you won’t know what you are missing!
Today, I am pleased to be able to tell you about three new features for WorkSpaces, each designed to make the service even more useful:
- Audio-In – You can now make and receive calls from your WorkSpace using popular communication tools such as Lync, Skype, and WebEx.
- High DPI Device Support – You can now take advantage of High DPI displays found on devices like the Surface Pro 4 tablet and the Lenovo Yoga laptop.
- Saved Registration Codes – You can now save multiple registration codes in the same client application.
Being able to make and to receive calls from your desktop can boost your productivity. Using the newest WorkSpaces clients for Windows and Mac, you can make and receive calls using popular communication tools like Lync, Skype, and WebEx. Simply connect an analog or USB audio headset to your local client device and start making calls! This functionality is enabled for all newly launched WorkSpaces; existing WorkSpaces may need a restart. With the launch of this feature, voice communication with headsets is available to you at no additional charge in all regions where WorkSpaces are available today.
When a WorkSpace is created using a custom image, the audio-in updates are applied during the provisioning process and will take some time. To avoid this, you (or your WorkSpaces administrator) can create a new custom image after the updates have been applied to an existing WorkSpace.
High DPI Devices
To support the increasing popularity of high DPI (Full HD, Ultra HD, and QHD+) displays, we added the ability to automatically scale the in-session experience of WorkSpaces to match your local DPI settings. This means that fonts and icon sizes will match your preferred settings on high DPI devices making the WorkSpaces experience more natural. Simply use the newest WorkSpaces clients for Windows and Mac and enjoy this enhancement immediately.
Saved Registration Codes
Many customers access multiple WorkSpaces spread across several directories and/or regions and would prefer not to have to copy and paste registration codes to make the switch. You can now save up to 10 registration codes within the client application, and switch between them with a couple of clicks. You can control all of this through the new Manage Registrations screen:
To learn more about Amazon WorkSpaces, visit the Amazon WorkSpaces page.— Jeff;
In the early days of AWS, customers were happy to simply learn about the cloud and its benefits. As they started to learn more, the conversation shifted. It went from “what is the cloud” to “what kinds of security does the cloud offer” to “”how can I use the cloud” over the course of just 6 or 7 years. As the industry begins to mature, enterprise and government customers are now interested in putting the cloud to use in a form that complies with applicable standards and recommendations.
For example, National Institute of Standards and Technology (NIST) Special Publication 800-53 (Security and Privacy Controls for Federal Information Systems and Organizations) defines a set of information and security controls that are designed to make systems more resilient to many different types of threats. This document is accompanied by a set of certifications, accreditations, and compliance processes.
New Compliance Offerings
In order to simplify the task of building a system that is in accord with compliance standards of this type, we will be publishing a series of AWS Enterprise Accelerator – Compliance Quick Starts. These documents and CloudFormation templates are designed to help Managed Service Organizations, cloud provisioning teams, developers, integrators, and information system security officers.
The new AWS Enterprise Accelerator – Compliance: Standardized Architecture for NIST 800-53 on the AWS Cloud is our first offering in this series!
The accelerator contains a set of nested CloudFormation templates. Deploying the top-level template takes about 30 minutes and creates all of the necessary AWS resources. The resources include three Virtual Private Clouds (VPCs)—Management, Development, and Production—suitable for running a multi-tier Linux-based application.
The template also creates the necessary IAM roles and custom policies, VPC security groups, and the like. It launches EC2 instances and sets up an encrypted, Multi-AZ MySQL database (using Amazon Relational Database Service (RDS)) in the Development and Production VPCs.
The architecture defined by this template makes use AWS best practices for security and availability including the use of a Multi-AZ architecture, isolation of instances between public and private subnets, monitoring & logging, database backup, and encryption.
You also have direct access to the templates. You can download them, customize them, and extract interesting elements for use in other projects.
You can also add the templates for this Quick Start to the AWS Service Catalog as portfolios or as products. This will allow you to institute a centrally managed model, and will help you to support consistent governance, security, and compliance.— Jeff;
For Episode 134 of the AWS podcast, I spoke with Bob Rogers, PhD, Chief Data Scientist for Big Data Solutions at Intel Corporation. We talked about how Bob entered the field of data science, how to get value from data science projects, and some misconceptions around big data. You can listen to the podcast to learn what skills are needed to have a career as a data scientist, and you can also hear Bob’s tips for those looking to become one. Hear what Intel is doing in the big data and analytics space – from the silicon chip to the cloud, and what big data holds for the future.
That’s the last of my 2015 recordings. We’ll be back with more episodes soon. Thanks for listening!— Jeff;
PS – Intel asked that we add the following disclaimer:
- Intel technologies’ features and benefits depend on system configuration and may require enabled hardware, software or service activation. Learn more at intel.com, or from the OEM or retailer.
- No computer system can be absolutely secure.
- Statements in this document that refer to Intel’s plans and expectations for the quarter, the year, and the future, are forward-looking statements that involve a number of risks and uncertainties. A detailed discussion of the factors that could affect Intel’s results and plans is included in Intel’s SEC filings, including the annual report on Form 10-K.
- Intel and the Intel logo are trademarks of Intel Corporation in the United States and/or other countries.