Amazon Elastic Block Store (EBS)
Introduction to Amazon EFS
EFS in Practice
Running Operations with the Snow Family
Storage for SAP on AWS
The course is part of this learning path
In this section of the AWS Certified: SAP on AWS Specialty learning path, we introduce you to the various Storage services currently available in AWS that are relevant to the PAS-C01 exam.
- Identify and describe the various Storage services available in AWS
- Understand how AWS Storage services can assist with large-scale data storage, migration, and transfer both into and out of AWS
- Describe hybrid cloud storage services and on-premises data backup solutions using AWS Storage services
- Identify storage options for SAP workloads on AWS
The AWS Certified: SAP on AWS Specialty certification has been designed for anyone who has experience managing and operating SAP workloads. Ideally you’ll also have some exposure to the design and implementation of SAP workloads on AWS, including migrating these workloads from on-premises environments. Many exam questions will require a solutions architect level of knowledge for many AWS services, including AWS Storage services. All of the AWS Cloud concepts introduced in this course will be explained and reinforced from the ground up.
Hello and welcome to this short lecture which will introduce you to the object level logging capabilities with your S3 buckets.
This feature is actually more closely related to the AWS CloudTrail service than S3 in a way, as it’s AWS CloudTrail that performs logging activities against Amazon S3 data events. These data events are specific API calls used in S3, such as
So what is CloudTrail? CloudTrail is a service that has a primary function to record and track all AWS API requests made. These API calls can be programmatic requests initiated from a user using an SDK, the AWS command-line interface, from within the AWS management console or even from a request made by another AWS service.
When an API request is initiated, AWS CloudTrail captures the request as an event and records this event within a log file which is then stored on S3. Each API call represents a new event within the log file. CloudTrail also records and associates other identifying metadata with all the events. For example, the identity of the caller, the timestamp of when the request was initiated and the source IP address.
We have a detailed course on AWS CloudTrail which an be found here which will provide a deep insight into the service and its full capabilities.
Capturing S3 data events can be configured in 2 ways: Firstly, if you want to capture data events for all or some of your S3 buckets, then you can configure this from within one of your Trails using the AWS CloudTrail console itself as shown here. Secondly, if it’s not already enabled via AWS CloudTrail for your bucket you can configure it at the bucket level using the Properties tab. Selecting the Object-level logging tile will present you with options to configure it.
As you can see, due to its integration with AWS CloudTrail you will be asked to select an existing trail from the same region to capture your S3 data events for this bucket. In this example, I have used my ‘Trail_Demo’ trail. You must also select which type of events you would like to capture, either just Read events or Write events, or both. Once you have made your selection, simply select Create and Object-level logging will be enabled and AWS CloudTrail will capture any S3 Data events associated with this bucket.
For more information on where your CloudTrail logs are stored and accessed, and how to interpret your CloudTrail logs, please see our existing course here.
Stuart has been working within the IT industry for two decades covering a huge range of topic areas and technologies, from data center and network infrastructure design, to cloud architecture and implementation.
To date, Stuart has created 150+ courses relating to Cloud reaching over 180,000 students, mostly within the AWS category and with a heavy focus on security and compliance.
Stuart is a member of the AWS Community Builders Program for his contributions towards AWS.
He is AWS certified and accredited in addition to being a published author covering topics across the AWS landscape.
In January 2016 Stuart was awarded ‘Expert of the Year Award 2015’ from Experts Exchange for his knowledge share within cloud services to the community.
Stuart enjoys writing about cloud technologies and you will find many of his articles within our blog pages.