Designing for high availability, fault tolerance and cost efficiency
High Availability in RDS
High Availability in Amazon Aurora
High Availability in DynamoDB
SAA-C02- Exam Prep
The course is part of this learning path
This section of the Solution Architect Associate learning path introduces you to the High Availability concepts and services relevant to the SAA-C02 exam. By the end of this section, you will be familiar with the design options available and know how to select and apply AWS services to meet specific availability scenarios relevant to the Solution Architect Associate exam.
- Learn the fundamentals of high availability, fault tolerance, and back up and disaster recovery
- Understand how a variety of Amazon services such as S3, Snowball, and Storage Gateway can be used for back up purposes
- Learn how to implement high availability practices in Amazon RDS, Amazon Aurora, and DynamoDB
Resources referenced within this lecture
Hello, and welcome to this lecture focusing on the AWS Snowball service. Essentially, this service is used to securely transfer large amounts of data and I'm talking up to petabyte scale here, in and out of AWS. Either from your on-premise data center to Amazon S3 or from Amazon S3 back to your data center using a physical appliance known as a snowball.
The snowball appliance comes as either a 50 terabyte or 80 terabyte storage device, depending on your region. Currently, the 50 terabyte version is only available within the U. S. regions. The appliance is dust, water, and tamper resistant and can even withstand a eight and a half G jolt from within it's own external shipping container.
And so it's been built to cope with a lot of stress conditions to ensure the durability of your data.
The snowball appliance has been designed to allow for high speed data transfer, thanks to a range of interfaces allowing you to select the most appropriate connection for your needs. Onboard the snowball appliance, the following I/O 10-gigabit interfaces are available, RJ45 using Cat6, SFP+ Copper and SFP+ Optical.
By default, all data transferred to the snowball appliance is automatically encrypted using 256-bit encryption keys generated from KMS, the key management service. Whilst on the topic of security, it also features end to end tracking using an E ink shipping label. This ensures that when the device leaves your premises, it is sent to the right AWS facility.
The appliance can also be tracked using the AWS Simple Notification Service with text messages or via the AWS Management Console.
From a compliance perspective, AWS Snowball is also HIPAA compliant allowing you to transfer protected health information in and out of S3.
When the transfer of data is complete via into S3 or into a customer's data center and the appliance is sent back to AWS. It is then the responsibility of AWS to ensure the data held in the Snowball Appliance is deleted and removed.
To control this process, AWS conforms to standards and guidelines set by NIST, the National Institute of Standard and Technology, to ensure this is performed and controlled and that all traces of data are removed from the media.
When sending or retrieving data, Snowball appliances can be aggregated together. For example, if you need to retrieve 400 terabytes of data from S3, then your data will be sent by five 80 terabyte Snowball appliances.
So from a disaster recovery perspective, when might you need to use AWS Snowball? Well, it all depends on how much data you need to get back from S3 to your own corporate data center and how quickly you can do that. On the other hand, how much data do you need to get into S3?
This'll depend on the connection you have to AWS from your data center. You may have direct connect connections, a VPN, or just an internet connection. And if you need to restore multiple petabytes of data, this could take weeks or even months to complete.
As a general rule, if your data retrieval will take longer than a week using your existing connection method, then you should consider using AWS Snowball.
Your global location will affect specific shipping times and so more information on this can be found using the link on the screen.
If you did decide to use AWS Snowball to retrieve your data in the event of a disaster, the process to use AWS Snowball is a fairly simple process. At a high level, this is how it looks.
- Firstly, you need to create an export job from within the AWS Management Console. Within this job, you can detect shipping details, the S3 bucket and the data to be exported, security mechanisms such as the KMS key for data encryption and also notifications.
- You will then receive delivery of your Snowball appliance.
- You can now connect the appliance to your local corporate network.
- Firstly, use the ports to connect the appliance to your network whilst it's powered off. Next, power on the device and the E Ink display will let you know that it's ready. You can then configure the network settings of the device such as the IP address to enable communications.
- From here, you are now ready to start transferring the data.
- To do this, you must first gain specific access credential for a manifest file through the Management Console which has to be downloaded. You must then install the Snowball Client software and you can now begin transferring data using the client software once authenticated with a manifest file.
- When the data transfer is complete, you can disconnect the Snowball Client. The appliance must then be returned to AWS using specified shipping carriers.
It's important to note that all Snowball appliances are the property of AWS. And the E Ink label will display the return address.
That now brings me to the end of this lecture while I looked at how and when you can use AWS Snowball to import or more importantly, export data, especially when required from a disaster recovery perspective.
Coming up next, I'll be explaining the different methods on how you can use the AWS Storage Gateway Service to transfer data between your data center and Amazon S3 for backup.
Stuart has been working within the IT industry for two decades covering a huge range of topic areas and technologies, from data center and network infrastructure design, to cloud architecture and implementation.
To date, Stuart has created 90+ courses relating to Cloud reaching over 140,000 students, mostly within the AWS category and with a heavy focus on security and compliance.
Stuart is a member of the AWS Community Builders Program for his contributions towards AWS.
He is AWS certified and accredited in addition to being a published author covering topics across the AWS landscape.
In January 2016 Stuart was awarded ‘Expert of the Year Award 2015’ from Experts Exchange for his knowledge share within cloud services to the community.
Stuart enjoys writing about cloud technologies and you will find many of his articles within our blog pages.