All About Amazon simple storage (S3)
Hii cloudy people
Welcome to Day 7 of our comprehensive 30-day aws learning path! In today’s session, we’ll be delving into the fascinating world of cloud Amazon S3 . By the end of this lesson, we’ll have a solid understanding of how it’s functions
Today, we’ll take a look at a brief amazon S3. Our goal is to focus on fundamentals then identify and learn the most commonly used services first and learn specialized services on a need to know basis.
What is Amazon S3?
· Amazon Simple Storage Service (S3) is a massively scalable storage service based on object storage technology
· Amazon S3 is an object storage service that offers industry-leading scalability, data availability, security, and performance.
· Store and protect any amount of data for a range of use cases, such as data lakes, websites, cloud-native applications, backups, archive, machine learning, and analytics.
· Amazon S3 is designed for 99.999999999% (11 9’s) of durability, and stores data for millions of customers all around the world.
· It provides a very high level of durability, with high availability and high performance. Data can be accessed from anywhere via the Internet, through the Amazon Console and the powerful S3 API.
S3 storage provides the following key features:
- Buckets — data is stored in buckets. Each bucket can store an unlimited amount of unstructured data.
- Elastic scalability — S3 has no storage limit. Individual objects can be up to 5TB in size.
- Flexible data structure — each object is identified using a unique key, and you can use metadata to flexibly organize data.
- Downloading data — easily share data with anyone inside or outside your organization and enable them to download data over the Internet.
- Permissions — assign permissions at the bucket or object level to ensure only authorized users can access data.
- APIs — the S3 API, provided both as REST and SOAP interfaces, has become an industry standard and is integrated with a large number of existing tools.
How Does S3 Storage Work?
Amazon S3 data is stored as objects. This approach enables highly scalable storage in the cloud. Objects can be placed on a variety of physical disk drives distributed throughout the data center. Amazon data centers use specialized hardware, software, and distributed file systems to provide true elastic scalability.
Amazon provides redundancy and version control using block storage methods. Data is automatically stored in multiple locations, distributed across multiple disks, and in some cases, multiple availability zones or regions. The Amazon S3 service periodically checks the integrity of the data by checking its control hash value. If data corruption is detected, redundant data is used to restore the object.
S3 lets you manage your data via the Amazon Console and the S3 API.
Buckets
Buckets are logical containers in which data is stored. S3 provides unlimited scalability, and there is no official limit on the amount of data and number of objects you can store in an S3 bucket. The size limit for objects stored in a bucket is 5 TB.
An S3 bucket name must be unique across all S3 users, because the bucket namespace is shared across all AWS accounts.
Keys
When you upload an object to a bucket, the object gets a unique key. The key is a string that mimics a directory hierarchy. Once you know the key, you can access the object in the bucket.
The bucket name, key, and version ID uniquely identify every object in S3. S3 provides two URL structures you can use to directly access an object:
{BUCKET-NAME}. s3.amazonaws.com/{OBJECT-KEY}
s3.amazon.aws.com/{BUCKET-NAME}/{OBJECT-KEY}
AWS Regions
Amazon has data centers in 24 geographical regions. To reduce network latency and minimize costs, store your data in the region closest to its users.
Unless you manually migrate your data, data stored in a specific AWS Region will never leave that region’s data center. AWS Regions are separated from each other to provide fault tolerance and reliability.
Each region is made up of at least three availability zones, which are separated, independent data centers. Data is replicated across availability zones to protect against outage of equipment in a specific data center, or disasters like fires, hurricanes and floods.
Amazon S3 Storage Classes
S3 provides storage tiers, also called storage classes, which can be applied at the bucket or object level. S3 also provides lifecycle policies you can use to automatically move objects between tiers, based on rules or thresholds you define.
The main storage classes are:
- Standard — for frequently accessed data
- Standard-IA — standard infrequent access
- One Zone-IA — one-zone infrequent access
- Intelligent-Tiering — automatically moves data to the most appropriate tier
Below we expand on the more commonly used classes.
Amazon S3 Standard
The S3 standard tier provides:
- Durability of 99.999999999% by replicating objects to multiple Availability Zones
- 99.99% availability backed by Service Level Agreement (SLA)
- Built-in SSL encryption for all data (both in transit and at rest)
Amazon S3 Standard-Infrequent Access
The S3 Standard-IA tier is for infrequently accessed data. It has a lower cost per GB/month, compared to the Standard tier, but charges a retrieval fee. The S3 Standard-IA tier provides:
- The same performance and latency as the Standard tier
- The same durability — 99.999999999% across multiple Availability Zones
- 99.9% availability backed by SLA
S3 Storage Archive
S3 provides Glacier and Deep Archive, storage classes intended for archived data that is accessed very infrequently. Cost per GB/month is lower than S3 Standard-IA.
- S3 Glacier — data must be stored for at least 90 days and can be restored within 1–5 minutes, with expedited retrieval.
- S3 Glacier Deep Archive — data must be stored for at least 180 days, and can be retrieved within 12 hours. There is a discount on bulk data retrieval, which takes up to 48 hours.
S3 Storage Q&A
How Much Data Can I Store in Amazon S3?
You can store an unlimited amount of data in Amazon S3. Other storage related limits include:
- Individual objects limited to 5TB
- Upload up to 5GB in one PUT operation
- For objects larger than 100MB Amazon recommends using Multiple Upload
How is Amazon S3 Data Organized?
Amazon S3 is an object store. Each object has a unique key that can be used to retrieve it later. You can define any string as a key, and keys can be used to create a hierarchy, for example by including a directory structure in the key. Another option is to organize objects using metadata, using S3 Object Tagging.
How Reliable is Amazon S3?
Amazon S3 provides 11 nines (99.999999999%) durability. With regard to availability, S3 guarantees:
- 99.99% availability for Standard storage class
- 99.9% availability for Standard-IA, Glacier and Deep Archive
- 99.5% availability for One Zone-IA
Thanks for reading
I hope found this useful?
✅Reshare this with others so that it helps more people
✅Tag your colleagues and friends who you think will need this