ExamGecko
Home Home / Amazon / DVA-C01

Amazon DVA-C01 Practice Test - Questions Answers, Page 14

Question list
Search
Search

List of questions

Search

Related questions











A company has three different environments: Development, QA, and Production. The company wants to deploy its code first in the Development environment, then QA, and then Production. Which AWS service can be used to meet this requirement?

A.
Use AWS CodeCommit to create multiple repositories to deploy the application.
A.
Use AWS CodeCommit to create multiple repositories to deploy the application.
Answers
B.
Use AWS CodeBuild to create, configure, and deploy multiple build application projects.
B.
Use AWS CodeBuild to create, configure, and deploy multiple build application projects.
Answers
C.
Use AWS Data Pipeline to create multiple data pipeline provisions to deploy the application.
C.
Use AWS Data Pipeline to create multiple data pipeline provisions to deploy the application.
Answers
D.
Use AWS CodeDeploy to create multiple deployment groups.
D.
Use AWS CodeDeploy to create multiple deployment groups.
Answers
Suggested answer: D

Explanation:

https://docs.aws.amazon.com/codedeploy/latest/userguide/deployment-groups.html"You can associate more than one deployment group with an application in CodeDeploy. This makes it possible to deploy an application revision to different sets of instances at different times. For example, you might use one deployment group to deploy an application revision to a set of instances tagged Test where you ensure the quality of the code. Next, you deploy the same application revision to a deployment group with instances tagged Staging for additional verification. Finally, when you are ready to release the latest application to customers, you deploy to a deployment group that includes instances tagged Production. "

A company is creating an application that will require users to access AWS services and allow them to reset their own passwords. Which of the following would allow the company to manage users and authorization while allowing users to reset their own passwords?

A.
Amazon Cognito identify pools and AWS STS
A.
Amazon Cognito identify pools and AWS STS
Answers
B.
Amazon Cognito identity pools and AWS IAM
B.
Amazon Cognito identity pools and AWS IAM
Answers
C.
Amazon Cognito user pools and AWS KMS
C.
Amazon Cognito user pools and AWS KMS
Answers
D.
Amazon Cognito user pools and identity pools
D.
Amazon Cognito user pools and identity pools
Answers
Suggested answer: D

Explanation:

https://serverless-stack.com/chapters/cognito-user-pool-vs-identity-pool.html

An application that runs on an Amazon EC2 instance needs to access and make API calls to multiple AWS services. What is the MOST secure way to provide access to the AWS services with MINIMAL management overhead?

A.
Use AWS KMS to store and retrieve credentials.
A.
Use AWS KMS to store and retrieve credentials.
Answers
B.
Use EC2 instance profiles.
B.
Use EC2 instance profiles.
Answers
C.
Use AWS root user to make requests to the application.
C.
Use AWS root user to make requests to the application.
Answers
D.
Store and retrieve credentials from AWS CodeCommit.
D.
Store and retrieve credentials from AWS CodeCommit.
Answers
Suggested answer: B

Explanation:

https://docs.aws.amazon.com/IAM/latest/UserGuide/id_roles_use_switch-role-ec2.html

In a multi-container Docker environment in AWS Elastic Beanstalk, what is required to configure container instances in the environment?

A.
An Amazon ECS task definition
A.
An Amazon ECS task definition
Answers
B.
An Amazon ECS cluster
B.
An Amazon ECS cluster
Answers
C.
A Docker in an application package
C.
A Docker in an application package
Answers
D.
A CLI for Elastic Beanstalk
D.
A CLI for Elastic Beanstalk
Answers
Suggested answer: A

Explanation:

Explanation:

Reference:

https://docs.aws.amazon.com/elasticbeanstalk/latest/dg/create_deploy_docker_ecs.html

A Development team currently supports an application that uses an in-memory store to save accumulated game results. Individual results are stored in a database. As part of migrating to AWS, the team needs to use automatic scaling. The team knows this will yield inconsistent results.

Where should the team store these accumulated game results to BEST allow for consistent results without impacting performance?

A.
Amazon S3
A.
Amazon S3
Answers
B.
Amazon RDS
B.
Amazon RDS
Answers
C.
Amazon ElastiCache
C.
Amazon ElastiCache
Answers
D.
Amazon Kinesis
D.
Amazon Kinesis
Answers
Suggested answer: C

A company has a multi-tiered web application on AWS. During a recent spike in traffic, one of the primary relational databases on Amazon RDS could not serve all the traffic. Some read queries for repeatedly accessed items failed, so users received error messages.

What can be done to minimize the impact on database read queries MOST efficiently during future traffic spikes?

A.
Use Amazon S3 to cache database query results.
A.
Use Amazon S3 to cache database query results.
Answers
B.
Use Amazon RDS as a custom origin for Amazon CloudFront.
B.
Use Amazon RDS as a custom origin for Amazon CloudFront.
Answers
C.
Use local storage and memory on Amazon EC2 instances to cache data.
C.
Use local storage and memory on Amazon EC2 instances to cache data.
Answers
D.
Use Amazon ElastiCache in front of the primary database to cache data.
D.
Use Amazon ElastiCache in front of the primary database to cache data.
Answers
Suggested answer: D

A company has a website that is developed in PHP and WordPress and is launched using AWS Elastic Beanstalk. There is a new version of the website that needs to be deployed in the Elastic Beanstalk environment. The company cannot tolerate having the website offline if an update fails.

Deployments must have minimal impact and rollback as soon as possible.

What deployment method should be used?

A.
All at once
A.
All at once
Answers
B.
Rolling
B.
Rolling
Answers
C.
Snapshots
C.
Snapshots
Answers
D.
Immutable
D.
Immutable
Answers
Suggested answer: D

Explanation:

https://docs.aws.amazon.com/elasticbeanstalk/latest/dg/environmentmgmt-updatesimmutable.html

A company maintains an application responsible for processing several thousand external callbacks each day. The company’s System administrators want to know how many callbacks are being received on a rolling basis, and they want this data available for 10 days. The company also wants the ability to issue automated alerts if the number of callbacks exceeds the defined thresholds. What is the MOST cost-effective way to address the need to track and alert on these statistics?

A.
Push callback data to an Amazon RDS database that can be queried to show historical data and to alert on exceeded thresholds.
A.
Push callback data to an Amazon RDS database that can be queried to show historical data and to alert on exceeded thresholds.
Answers
B.
Push callback data to AWS X-Ray and use AWS Lambda to query, display, and alert on exceeded thresholds.
B.
Push callback data to AWS X-Ray and use AWS Lambda to query, display, and alert on exceeded thresholds.
Answers
C.
Push callback data to Amazon Kinesis Data Streams and invoke an AWS Lambda function that stores data in Amazon DynamoDB and sends the required alerts.
C.
Push callback data to Amazon Kinesis Data Streams and invoke an AWS Lambda function that stores data in Amazon DynamoDB and sends the required alerts.
Answers
D.
Push callback data to Amazon CloudWatch as a custom metric and use the CloudWatch alerting mechanisms to alert System Administrators.
D.
Push callback data to Amazon CloudWatch as a custom metric and use the CloudWatch alerting mechanisms to alert System Administrators.
Answers
Suggested answer: D

Developer is creating an AWS Lambda function to process a stream of data from an Amazon Kinesis Data Stream. When the Lambda function parses the data and encounters a missing field, it exits the function with an error. The function is generating duplicate records from the Kinesis stream. When the Developer looks at the stream output without the Lambda function, there are no duplicate records. What is the reason for the duplicates?

A.
The Lambda function did not advance the Kinesis stream pointer to the next record after the error.
A.
The Lambda function did not advance the Kinesis stream pointer to the next record after the error.
Answers
B.
The Lambda event source used asynchronous invocation, resulting in duplicate records.
B.
The Lambda event source used asynchronous invocation, resulting in duplicate records.
Answers
C.
The Lambda function did not handle the error, and the Lambda service attempted to reprocess the data.
C.
The Lambda function did not handle the error, and the Lambda service attempted to reprocess the data.
Answers
D.
The Lambda function is not keeping up with the amount of data coming from the stream.
D.
The Lambda function is not keeping up with the amount of data coming from the stream.
Answers
Suggested answer: C

Explanation:

https://docs.aws.amazon.com/lambda/latest/dg/with-kinesis.html

A Developer must build an application that uses Amazon DynamoDB. The requirements state that items being stored in the DynamoDB table will be 7KB in size and that reads must be strongly consistent. The maximum read rate is 3 items per second, and the maximum write rate is 10 items per second.

How should the Developer size the DynamoDB table to meet these requirements?

A.
Read: 3 read capacity unitsWrite: 70 write capacity units
A.
Read: 3 read capacity unitsWrite: 70 write capacity units
Answers
B.
Read: 6 read capacity unitsWrite: 70 write capacity units
B.
Read: 6 read capacity unitsWrite: 70 write capacity units
Answers
C.
Read: 6 read capacity unitsWrite: 10 write capacity units
C.
Read: 6 read capacity unitsWrite: 10 write capacity units
Answers
D.
Read: 3 read capacity unitsWrite: 10 write capacity units
D.
Read: 3 read capacity unitsWrite: 10 write capacity units
Answers
Suggested answer: B

Explanation:

Explanation:

https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Limits.html

Total 608 questions
Go to page: of 61