In this course, we provide an overview of storage options for SAP environments running on AWS.
- A greater understanding of the various storage offerings available when architecting SAP workloads on AWS
- The use cases associated with each of the storage options and be able to describe the enhanced flexibility, durability, and security they provide
- Anyone responsible for implementing and managing SAP workloads on AWS
- Anyone looking to take the AWS Certified: SAP on AWS - Specialty certification exam
- You should have a basic understanding of AWS, including AWS storage services
- It would also be useful to have some experience provisioning Amazon EC2 instances and creating and configuring EBS volumes, S3 buckets, and EFS or FSx file systems
- For more information on all of these services, please check out this learning path and these courses:
Hello, and welcome to this lecture, where I will be discussing block storage using Amazon EBS–Elastic Block Storage–for SAP workloads on AWS. By the end of this lecture, you’ll understand how to use Amazon EBS with the EC2 instances that are part of your SAP deployments on AWS, along with the different types of EBS volumes that you can provision and when you should use them.
Amazon EBS provides raw, unformatted block storage volumes, which you can think of as being like traditional disk drives. EBS volumes can range in size from 1 GB all the way up to 16 TB. You can create file systems on top of EBS volumes, and multiple EBS volumes may be attached to a single EC2 instance; however, one EBS volume may only be attached to a single EC2 instance at a time.
EBS volumes may also be combined in any standard RAID configuration for additional redundancy, as long as the underlying operating system of the EC2 instance supports it. Now it’s worth pointing out that all EBS volumes are already replicated within a single availability zone to protect against the failure of a single underlying disk, giving you five nines, or 99.999% availability, per EBS volume. So again, you’ll be using EBS volumes as storage volumes for the EC2 instances that host your SAP applications and databases, where they can be used to store everything from the data residing in your SAP databases, to transaction and application log files, or even local backups.
Although you’ll also see that Amazon S3 may be better suited for long-term backup storage. Now unlike EC2 Instance Storage which is ephemeral, data stored on EBS volumes persists independently from the lifecycle of the EC2 instance to which it is attached. EBS volumes can also be backed up using snapshots, which are incremental, point-in-time backups that are stored in Amazon S3. You can even use EBS snapshots to create replicas of existing SAP systems for development or testing. And EBS volumes as well as their snapshots may also be encrypted for additional security, with no additional cost or performance impacts.
Now when creating a new EBS volume, one of the first choices you’ll need to make is which volume type to use. So broadly speaking, EBS volume types fall into one of two categories: solid state drives, or SSDs, or magnetic hard disk drives, or HDDs. And these are no different than the SSDs or magnetic disks you might use in your on-premises servers, with the same reliability, cost, and performance tradeoff considerations.
So SSDs are flash storage with no moving parts and because of this, they perform much faster and are much more reliable than traditional magnetic storage. This makes them ideally suited for SAP transactional workloads and instance boot volumes, where read and write speed is most important. Valid use cases for SSDs include file systems for SAP applications, SAP database log and data files, and even SAP local backups. Now we measure the performance of our volumes using units referred to as input/output operations per second, or IOPS. And in AWS, your choices for solid-state storage are the general purpose, or gp2 and gp3 volume types, or the provisioned IOPS, or io1 and io2 volume types.
So both general-purpose and provisioned IOPS SSD storage offer single-digit millisecond latency. But while gp2 and gp3 volumes can offer up to 16,000 IOPS per volume, both the io1 and io2 volume types allow you to scale up to 64,000 IOPS, which may be necessary if your SAP database has sustained I/O-intensive throughput requirements. But as you might expect, this performance will come with a higher cost.
So how can you determine which volume type is best suited for your SAP workloads? Well, if you’ve calculated your workload’s SAP Application Performance Standard, or SAPS benchmark using something like the SAP Quick Sizer tool, you can estimate the total number of IOPS you’ll need by assuming a value of 60-90% of the Database SAPS for your workload. Now for your Online Transaction Processing or OLTP workloads, this factor will be closer to 60% of your Database SAPS and for your Online Analytical Processing, or OLAP workloads, it may be closer to 90%. So for instance, if you’re running an OLTP workload such as SAP Business Suite with a total of 10,000 Database SAPS, you can expect to need around 6,000 IOPS for your database storage volume, which can easily be achieved with a gp2 storage volume. But if you’re going to need more than 16,000 IOPS, you’ll need to use a Provisioned IOPS volume instead.
So generally speaking, you’re going to want to always start by using a General Purpose gp2 volume and seeing if that is capable of meeting your performance requirements. Gp2 volumes are typically sufficient for most SAP HANA workloads. But if they aren’t, then you can quickly and easily change to using a Provisioned IOPS volume instead. So for instance, you could start with a gp2 volume and upgrade it to an io1 or io2 volume later on if the gp2 volume doesn’t meet your performance requirements. Or you could use a gp3 volume type instead, which allows you to provision your IOPS and throughput independently of the storage capacity of your volume. And this could be useful if you have highly transactional SAP workloads that don’t require a lot of additional storage space.
So we’ve seen how SSD-backed EBS volume types are useful for boot volumes as well as critical applications and transactional databases that have low latency and high throughput requirements. But for streaming workloads, or for large volumes of data that are not frequently accessed, HDD-backed volumes are a much more cost-effective alternative. And again, HDDs are your typical magnetic storage disks that have moving mechanical parts.
Now for streaming workloads, a Throughput Optimized st1 volume offers a maximum throughput of up to 500 MB per second at roughly half the cost of a similarly-sized gp2 or gp3 volume. And finally, your lowest cost storage option is the Cold HD, or sc1 storage volume. And just like our other volumes, these sc1 volumes can be up to 16 TB in size and still offer a maximum throughput of up to 250 MB per second, but at about one-third of the cost of an equivalent st1 volume. So while an st1 volume might be useful for streaming workloads like log processing, an sc1 volume is better suited for things like long-term data archives, or any other data that is large, but not frequently accessed.
Now keep in mind that you’re free to mix these different volume types within your overall SAP solution architecture. So it’s entirely valid to have, for instance, a gp2 boot volume and a separate io1 storage volume attached to the same EC2 instance for your SAP database in a way that allows you to effectively balance both cost and performance considerations. You could even leverage the ephemeral EC2 instance store for things like temporary cache files that require the absolute fastest possible read and write speeds, but do not need to be persisted if the associated EC2 instance is terminated. And as I previously mentioned, it is possible to transition between different volume types over time as requirements change. You can also leverage Amazon CloudWatch, where EBS sends data points regarding burst balance, queue length, and read and write IOPS, to monitor the performance of your EBS volumes over time. For more information on infrastructure monitoring using Amazon CloudWatch, check out this course:
Danny has over 20 years of IT experience as a software developer, cloud engineer, and technical trainer. After attending a conference on cloud computing in 2009, he knew he wanted to build his career around what was still a very new, emerging technology at the time — and share this transformational knowledge with others. He has spoken to IT professional audiences at local, regional, and national user groups and conferences. He has delivered in-person classroom and virtual training, interactive webinars, and authored video training courses covering many different technologies, including Amazon Web Services. He currently has six active AWS certifications, including certifications at the Professional and Specialty level.