AWS Aurora database engine is fully released, it’s time to see how it runs in the real world.
For some time, AWS has offered managed relational databases (RDS) for MySQL, SQLServer, Oracle, and Postgres deployments. Now, after some months in preview, Amazon has officially launched the full version of their own AWS Aurora database.
In general, Amazon’s RDS handles administration, backups, and snapshots, leaving you to worry about your data. You can easily scale your compute and storage resources to fit your evolving needs. As with all AWS managed services, RDS is highly available and you can use their MultiAZ feature to synchronously replicate data to a standby instance in a different availability zone.
You can access these databases through the AWS Management Console, the command line tool, and API calls.
AWS Aurora was built to deliver significantly improved parallel processing and concurrent I/O operations. In traditional database engine architectures (like Mysql, MSSQL, and Oracle), all layers of data functionality – like SQL, transactions, caching, and logging – reside in single box.
But when you provision Amazon Aurora, logging and caching are moved into a “multi tenant, scale-out database-optimised storage service” that’s deeply integrated with other AWS compute and storage services. Besides allowing you to dramatically scale without nearly the overhead, you can restart the database engine without losing the cache.
Why you might want to use AWS Aurora
- Amazon Aurora is compatible with MySQL 5.6. That means the applications, tools, and drivers you’re already using with MySQL can be used.
- You can launch a new Aurora database from an RDS MySQL DB snapshot.
- 5x increase in performance when compared to MySQL.
- You can scale your database up to 32 virtual CPU’s (vCPU) and 244 GB memory. You can optionally add up to fifteen read replicas to increase read capacity.
- Aurora is designed to detect database crashes and restart without the need for manual crash recovery.
- Highly available: the db is replicated six ways across three Availability Zones.
- Automatic storage scaling up to 64 TB.
- AWS Aurora uses SSL (AES-256) encryption to secure data in transit.
- The Aurora database engine is able to provide 100,000 writes and 500,000 reads per second.
Comparison: Mysql vs AWS Aurora
Here’s Amazon’s own feature-by-feature comparison of RDS MySQL and Aurora:
Amazon Aurora Pricing
- Database instances: AWS RDS instances are available in 5 sizes starting from 2vCPU to 32vCPU, using between 15.25 to 244 GB of memory:
- Storage: $0.100 per GB-month.
- I/O Rate: $0.200 per 1 million requests.
As of now, Amazon Aurora is only available for instances starting with dbr3.large (2vCPU/15GB). This means that the option is effectively not available for smaller instances. Aurora only supports InnoDB, and when you migrate from MySQL, MYISAM tables are automatically converted to InnoDB.
Keeping in all this in mind, while Aurora is now publicly available, you’re certainly free to start playing around with it as part of a test or POC environment. As you gain experience, you’ll see some of the platform’s advantages and limitations for yourself. You might also want to browse through the testimonials of other AWS customers who are already deeply involved in their own explorations.
Two New EC2 Instance Types Announced at AWS re:Invent 2018 – Monday Night Live
Let’s look at what benefits these two new EC2 instance types offer and how these two new instances could be of benefit to you. Both of the new instance types are built on the AWS Nitro System. The AWS Nitro System improves the performance of processing in virtualized environments by...
Google Cloud Certification: Preparation and Prerequisites
Google Cloud Platform (GCP) has evolved from being a niche player to a serious competitor to Amazon Web Services and Microsoft Azure. In 2018, research firm Gartner placed Google in the Leaders quadrant in its Magic Quadrant for Cloud Infrastructure as a Service for the first time. In t...
Understanding AWS VPC Egress Filtering Methods
Security in AWS is governed by a shared responsibility model where both vendor and subscriber have various operational responsibilities. AWS assumes responsibility for the underlying infrastructure, hardware, virtualization layer, facilities, and staff while the subscriber organization ...
S3 FTP: Build a Reliable and Inexpensive FTP Server Using Amazon’s S3
Is it possible to create an S3 FTP file backup/transfer solution, minimizing associated file storage and capacity planning administration headache?FTP (File Transfer Protocol) is a fast and convenient way to transfer large files over the Internet. You might, at some point, have conf...
Microservices Architecture: Advantages and Drawbacks
Microservices are a way of breaking large software projects into loosely coupled modules, which communicate with each other through simple Application Programming Interfaces (APIs).Microservices have become increasingly popular over the past few years. The modular architectural style,...
What Are Best Practices for Tagging AWS Resources?
There are many use cases for tags, but what are the best practices for tagging AWS resources? In order for your organization to effectively manage resources (and your monthly AWS bill), you need to implement and adopt a thoughtful tagging strategy that makes sense for your business. The...
How to Optimize Amazon S3 Performance
Amazon S3 is the most common storage options for many organizations, being object storage it is used for a wide variety of data types, from the smallest objects to huge datasets. All in all, Amazon S3 is a great service to store a wide scope of data types in a highly available and resil...
How to Optimize Cloud Costs with Spot Instances: New on Cloud Academy
One of the main promises of cloud computing is access to nearly endless capacity. However, it doesn’t come cheap. With the introduction of Spot Instances for Amazon Web Services’ Elastic Compute Cloud (AWS EC2) in 2009, spot instances have been a way for major cloud providers to sell sp...
What are the Benefits of Machine Learning in the Cloud?
A Comparison of Machine Learning Services on AWS, Azure, and Google CloudArtificial intelligence and machine learning are steadily making their way into enterprise applications in areas such as customer support, fraud detection, and business intelligence. There is every reason to beli...
How to Use AWS CLI
The AWS Command Line Interface (CLI) is for managing your AWS services from a terminal session on your own client, allowing you to control and configure multiple AWS services.So you’ve been using AWS for awhile and finally feel comfortable clicking your way through all the services....
AWS Summit Chicago: New AWS Features Announced
Thousands of cloud practitioners descended on Chicago’s McCormick Place West last week to hear the latest updates around Amazon Web Services (AWS). While a typical hot and humid summer made its presence known outside, attendees inside basked in the comfort of air conditioning to hone th...
From Monolith to Serverless – The Evolving Cloudscape of Compute
Containers can help fragment monoliths into logical, easier to use workloads. The AWS Summit New York was held on July 17 and Cloud Academy sponsored my trip to the event. As someone who covers enterprise cloud technologies and services, the recent Amazon Web Services event was an insig...