AWS Interview Questions 5

  1. When you need to move data over long distances using the internet, for instance across countries or continents to your Amazon S3 bucket, which method or service will you use?
    1. Amazon Glacier
    2. Amazon CloudFront
    3. Amazon Transfer Acceleration
    4. Amazon Snowball

Answer C.

Explanation: You would not use Snowball, because for now, the snowball service does not support cross region data transfer, and since, we are transferring across countries, Snowball cannot be used. Transfer Acceleration shall be the right choice here as it throttles your data transfer with the use of optimized network paths and Amazon’s content delivery network upto 300% compared to normal data transfer speed.

22. How can you speed up data transfer in Snowball?

READ ALSO
AWS Interview Questions 1

The data transfer can be increased in the following way:

  • By performing multiple copy operations at one time i.e. if the workstation is powerful enough, you can initiate multiple cp commands each from different terminals, on the same Snowball device.

 

    • Copying from multiple workstations to the same snowball.
    • Transferring large files or by creating a batch of small file, this will reduce the encryption overhead.
    • Eliminating unnecessary hops i.e. make a setup where the source machine(s) and the snowball are the only machines active on the switch being used, this can hugely improve performance.
  1. If you want to launch Amazon Elastic Compute Cloud (EC2) instances and assign each instance a predetermined private IP address you should:
  2. Launch the instance from a private Amazon Machine Image (AMI).
  3. Assign a group of sequential Elastic IP address to the instances.
  4. Launch the instances in the Amazon Virtual Private Cloud (VPC).
  5. Launch the instances in a Placement Group.
READ ALSO
ESB Amazon Infrastructure (AWS)

Answer C.

Explanation: The best way of connecting to your cloud resources (for ex- ec2 instances) from your own data center (for eg- private cloud) is a VPC. Once you connect your datacenter to the VPC in which your instances are present, each instance is assigned a private IP address which can be accessed from your datacenter. Hence, you can access your public cloud resources, as if they were on your own network.

24. Can I connect my corporate datacenter to the Amazon Cloud?

Yes, you can do this by establishing a VPN(Virtual Private Network) connection between your company’s network and your VPC (Virtual Private Cloud), this will allow you to interact with your EC2 instances as if they were within your existing network.

  1. Is it possible to change the private IP addresses of an EC2 while it is running/stopped in a VPC?
READ ALSO
Interview Questions

Primary private IP address is attached with the instance throughout its lifetime and cannot be changed, however secondary private addresses can be unassigned, assigned or moved between interfaces or instances at any point.

  1. Why do you make subnets?
    1. Because there is a shortage of networks
    2. To efficiently utilize networks that have a large no. of hosts.
    3. Because there is a shortage of hosts.
    4. To efficiently utilize networks that have a small no. of hosts.

Answer B.

 

Explanation: If there is a network which has a large no. of hosts, managing all these hosts can be a tedious job. Therefore we divide this network into subnets (sub-networks) so that managing these hosts becomes simpler.

  1. Which of the following is true?
    1. You can attach multiple route tables to a subnet
    2. You can attach multiple subnets to a route table
    3. Both A and B
    4. None of these.
READ ALSO
What is DevOps? Introduction to DevOps 6

Answer B.

Explanation: Route Tables are used to route network packets, therefore in a subnet having multiple route tables will lead to confusion as to where the packet has to go. Therefore, there is only one route table in a subnet, and since a route table can have any no. of records or information, hence attaching multiple subnets to a route table is possible.

  1. In CloudFront what happens when content is NOT present at an Edge location and a request is made to it?
    1. An Error “404 not found” is returned
    2. CloudFront delivers the content directly from the origin server and stores it in the cache of the edge location
    3. The request is kept on hold till content is delivered to the edge location
    4. The request is routed to the next closest edge location
READ ALSO
AWS Interview Questions 3

Answer B.

Explanation: CloudFront is a content delivery system, which caches data to the nearest edge location from the user, to reduce latency. If data is not present at an edge location, the first time the data may get transferred from the original server, but from the next time, it will be served from the cached edge.

  1. If I’m using Amazon CloudFront, can I use Direct Connect to transfer objects from my own data center?

Yes. Amazon CloudFront supports custom origins including origins from outside of AWS. With AWS Direct Connect, you will be charged with the respective data transfer rates.

30. If my AWS Direct Connect fails, will I lose my connectivity?

READ ALSO
Ansible Interview Questions and Answers 2

If a backup AWS Direct connect has been configured, in the event of a failure it will switch over to the second one. It is recommended to enable Bidirectional Forwarding Detection (BFD) when configuring

 

your connections to ensure faster detection and failover. On the other hand, if you have configured a backup IPsec VPN connection instead, all VPC traffic will failover to the backup VPN connection automatically. Traffic to/from public resources such as Amazon S3 will be routed over the Internet. If you do not have a backup AWS Direct Connect link or a IPsec VPN link, then Amazon VPC traffic will be dropped in the event of a failure.

  1. If I launch a standby RDS instance, will it be in the same Availability Zone as my primary?
    1. Only for Oracle RDS types
    2. Yes
    3. Only if it is configured at launch
    4. No
READ ALSO
Raster Vector Data Analysis ~ Hiking Path Finder

Answer D.

Explanation: No, since the purpose of having a standby instance is to avoid an infrastructure failure (if it happens), therefore the standby instance is stored in a different availability zone, which is a physically different independent infrastructure.

  1. When would I prefer Provisioned IOPS over Standard RDS storage?
    1. If you have batch-oriented workloads
    2. If you use production online transaction processing (OLTP) workloads.
    3. If you have workloads that are not sensitive to consistent performance
    4. All of the above

Answer A.

Explanation: Provisioned IOPS deliver high IO rates but on the other hand it is expensive as well. Batch processing workloads do not require manual intervention they enable full utilization of systems, therefore a provisioned IOPS will be preferred for batch oriented workload.

READ ALSO
Using R on Amazon EC2 under the Free Usage Tier

33. How is Amazon RDS, DynamoDB and Redshift different?

 Amazon RDS is a database management service for relational databases, it manages patching, upgrading, backing up of data etc. of databases for you without your intervention. RDS is a Db management service for structured data only.

  • DynamoDB, on the other hand, is a NoSQL database service, NoSQL deals with unstructured data.
  • Redshift, is an entirely different service, it is a data warehouse product and is used in data analysis.

 

  1. If I am running my DB Instance as a Multi-AZ deployment, can I use the standby DB Instance for read or write operations along with primary DB instance?
    1. Yes
    2. Only with MySQL based RDS
    3. Only for Oracle RDS instances
    4. No
READ ALSO
BD & BA – ISfBD Practical AWS

Answer D.

Explanation: No, Standby DB instance cannot be used with primary DB instance in parallel, as the former is solely used for standby purposes, it cannot be used unless the primary instance goes down.

  1. Your company’s branch offices are all over the world, they use a software with a multi-regional deployment on AWS, they use MySQL 5.6 for data persistence.

The task is to run an hourly batch process and read data from every region to compute cross-regional reports which will be distributed to all the branches. This should be done in the shortest time possible. How will you build the DB architecture in order to meet the requirements?

  1. For each regional deployment, use RDS MySQL with a master in the region and a read replica in the HQ region
  2. For each regional deployment, use MySQL on EC2 with a master in the region and send hourly EBS snapshots to the HQ region
  3. For each regional deployment, use RDS MySQL with a master in the region and send hourly RDS snapshots to the HQ region
  4. For each regional deployment, use MySQL on EC2 with a master in the region and use S3 to copy data files hourly to the HQ region
READ ALSO
alexa-skills-kit

Answer A.

Explanation: For this we will take an RDS instance as a master, because it will manage our database for us and since we have to read from every region, we’ll put a read replica of this instance in every region where the data has to be read from. Option C is not correct since putting a read replica would be more efficient than putting a snapshot, a read replica can be promoted if needed to an independent DB instance, but with a Db snapshot it becomes mandatory to launch a separate DB Instance.

36. Can I run more than one DB instance for Amazon RDS for free?

READ ALSO
Jenkins Interview Questions

Yes. You can run more than one Single-AZ Micro database instance, that too for free! However, any use exceeding 750 instance hours, across all Amazon RDS Single-AZ Micro DB instances, across all eligible database engines and regions, will be billed at standard Amazon RDS prices. For example: if

 

you run two Single-AZ Micro DB instances for 400 hours each in a single month, you will accumulate 800 instance hours of usage, of which 750 hours will be free. You will be billed for the remaining 50 hours at the standard Amazon RDS price.

  1. Which AWS services will you use to collect and process e-commerce data for near real-time analysis?
    1. Amazon ElastiCache
    2. Amazon DynamoDB
    3. Amazon Redshift
    4. Amazon Elastic MapReduce
READ ALSO
Top 10 DevOps Tools

Answer B,C.

Explanation: DynamoDB is a fully managed NoSQL database service. DynamoDB, therefore can be fed any type of unstructured data, which can be data from e-commerce websites as well, and later, an analysis can be done on them using Amazon Redshift. We are not using Elastic MapReduce, since a near real time analyses is needed.

  1. Can I retrieve only a specific element of the data, if I have a nested JSON data in DynamoDB?

Yes. When using the GetItem, BatchGetItem, Query or Scan APIs, you can define a Projection Expression to determine which attributes should be retrieved from the table. Those attributes can include scalars, sets, or elements of a JSON document.

  1. A company is deploying a new two-tier web application in AWS. The company has limited staff and requires high availability, and the application requires complex queries and table joins.
READ ALSO
DevOps Interview Questions

Which configuration provides the solution for the company’s requirements?

    1. MySQL Installed on two Amazon EC2 Instances in a single Availability Zone
    2. Amazon RDS for MySQL with Multi-AZ
    3. Amazon ElastiCache
    4. Amazon DynamoDB

Answer D.

Explanation: DynamoDB has the ability to scale more than RDS or any other relational database service, therefore DynamoDB would be the apt choice.

40. What happens to my backups and DB Snapshots if I delete my DB Instance?

 

When you delete a DB instance, you have an option of creating a final DB snapshot, if you do that you can restore your database from that snapshot. RDS retains this user-created DB snapshot along with all other manually created DB snapshots after the instance is deleted, also automated backups are deleted and only manually created DB Snapshots are retained.

READ ALSO
Getting started with amazon-ec2