Object Storage and Amazon S3
-
Amazon S3 stores objects in containers which are referred as buckets
-
We can set access controls specific to an individual bucket or an individual object(s)
-
Creating an S3 bucket:
- To create an s3 bucket we need to
- choose the region
- specify a bucket name
- AWS creates a virtual hosted URL
https://<bucket name>.s3.amazonaws.com https://<bucket name>.s3.<region name>.amazonaws.com- Regions:

- S3 urls are exposed using endpoints Refer Here
Region name Region URL Examples of S3 endpoints US East (N. Virgina) us-east-1 s3.amazonaws.com US West (Oregon) us-west-2 s3.us-west-2.amazonaws.com Asia Pacific (Mumbai) ap-south-1 s3.ap-south-1.amazonaws.com - Bucket names should be unique across AWS
- Lets create a s3 bucket and upload an image

- For Bucket naming rules Refer Here
- In S3 buckets we have objects and folders

- To create an s3 bucket we need to
-
When it comes to storing the data there are following concerns
- Running out of space => S3 has unlimited storage
- Durability:
- The durability of S3 is 99.99999999999 (elven 9’s) percentage.
- The above number means if we store 10000 objects with Amazon S3, we might loose one object every 10 million years.
- Availability: The availability is determined by which storage class you are using in AWS S3. The availability promised is also defined in SLA.
-
In AWS we can create up to 100 buckets per account and this can be increased to a maximum of 1000 buckets.
S3 Pricing
- We use S3 to store data. The pricing of S3 is two fold
- Paying for storage
- Paying for accessing
- To tune your bucket or objects according to the needs AWS has storage classes.
- Storage classes also impact availability

- In AWS we have the following storage classes
- Standard
- Infrequent access
- One-Zone Infrequent access
- Intelligent tier
- glacier
- Reduced Redundancy
