ExamGecko
Home Home / Amazon / BDS-C00

Amazon BDS-C00 Practice Test - Questions Answers, Page 8

Question list
Search
Search

List of questions

Search

Related questions











An Operations team continuously monitors the number of visitors to a website to identify any potential system problems. The number of website visitors varies throughout the day. The site is more popular in the middle of the day and less popular at night.

Which type of dashboard display would be the MOST useful to allow staff to quickly and correctly identify system problems?

A.
A vertical stacked bar chart showing today's website visitors and the historical average number of website visitors.
A.
A vertical stacked bar chart showing today's website visitors and the historical average number of website visitors.
Answers
B.
An overlay line chart showing today's website visitors at one-minute intervals and also the historical average number of website visitors.
B.
An overlay line chart showing today's website visitors at one-minute intervals and also the historical average number of website visitors.
Answers
C.
A single KPI metric showing the statistical variance between the current number of website visitors and the historical number of website visitors for thecurrent time of day.
C.
A single KPI metric showing the statistical variance between the current number of website visitors and the historical number of website visitors for thecurrent time of day.
Answers
D.
A scatter plot showing today's website visitors on the X-axis and the historical average number of website visitors on the Y-axis.
D.
A scatter plot showing today's website visitors on the X-axis and the historical average number of website visitors on the Y-axis.
Answers
Suggested answer: B

An organization would like to run analytics on their Elastic Load Balancing logs stored in Amazon S3 and join this data with other tables in Amazon S3. The users are currently using a BI tool connecting with JDBC and would like to keep using this BI tool.

Which solution would result in the LEAST operational overhead?

A.
Trigger a Lambda function when a new log file is added to the bucket to transform and load it into Amazon Redshift. Run the VACUUM command on theAmazon Redshift cluster every night.
A.
Trigger a Lambda function when a new log file is added to the bucket to transform and load it into Amazon Redshift. Run the VACUUM command on theAmazon Redshift cluster every night.
Answers
B.
Launch a long-running Amazon EMR cluster that continuously downloads and transforms new files from Amazon S3 into its HDFS storage. Use Presto toexpose the data through JDBC.
B.
Launch a long-running Amazon EMR cluster that continuously downloads and transforms new files from Amazon S3 into its HDFS storage. Use Presto toexpose the data through JDBC.
Answers
C.
Trigger a Lambda function when a new log file is added to the bucket to transform and move it to another bucket with an optimized data structure. UseAmazon Athena to query the optimized bucket.
C.
Trigger a Lambda function when a new log file is added to the bucket to transform and move it to another bucket with an optimized data structure. UseAmazon Athena to query the optimized bucket.
Answers
D.
Launch a transient Amazon EMR cluster every night that transforms new log files and loads them into Amazon Redshift.
D.
Launch a transient Amazon EMR cluster every night that transforms new log files and loads them into Amazon Redshift.
Answers
Suggested answer: C

An organization has added a clickstream to their website to analyze traffic. The website is sending each page request with the PutRecord API call to an Amazon Kinesis stream by using the page name as the partition key. During peak spikes in website traffic, a support engineer notices many ProvisionedThroughputExcededException events in the application logs.

What should be done to resolve the issue in the MOST cost-effective way?

A.
Create multiple Amazon Kinesis streams for page requests to increase the concurrency of the clickstream.
A.
Create multiple Amazon Kinesis streams for page requests to increase the concurrency of the clickstream.
Answers
B.
Increase the number of shards on the Kinesis stream to allow for more throughput to meet the peak spikes in traffic.
B.
Increase the number of shards on the Kinesis stream to allow for more throughput to meet the peak spikes in traffic.
Answers
C.
Modify the application to use on the Kinesis Producer Library to aggregate requests before sending them to the Kinesis stream.
C.
Modify the application to use on the Kinesis Producer Library to aggregate requests before sending them to the Kinesis stream.
Answers
D.
Attach more consumers to the Kinesis stream to process records in parallel, improving the performance on the stream.
D.
Attach more consumers to the Kinesis stream to process records in parallel, improving the performance on the stream.
Answers
Suggested answer: B

Explanation:

Reference: https://aws.amazon.com/kinesis/data-streams/faqs/

An organization currently runs a large Hadoop environment in their data center and is in the process of creating an alternative Hadoop environment on AWS, using Amazon EMR.

They generate around 20 TB of data on a monthly basis. Also on a monthly basis, files need to be grouped and copied to Amazon S3 to be used for the Amazon

EMR environment. They have multiple S3 buckets across AWS accounts to which data needs to be copied. There is a 10G AWS Direct Connect setup between their data center and AWS, and the network team has agreed to allocate

A.
Use an offline copy method, such as an AWS Snowball device, to copy and transfer data to Amazon S3.
A.
Use an offline copy method, such as an AWS Snowball device, to copy and transfer data to Amazon S3.
Answers
B.
Configure a multipart upload for Amazon S3 on AWS Java SDK to transfer data over AWS Direct Connect.
B.
Configure a multipart upload for Amazon S3 on AWS Java SDK to transfer data over AWS Direct Connect.
Answers
C.
Use Amazon S3 transfer acceleration capability to transfer data to Amazon S3 over AWS Direct Connect.
C.
Use Amazon S3 transfer acceleration capability to transfer data to Amazon S3 over AWS Direct Connect.
Answers
D.
Setup S3DistCop tool on the on-premises Hadoop environment to transfer data to Amazon S3 over AWS Direct Connect.
D.
Setup S3DistCop tool on the on-premises Hadoop environment to transfer data to Amazon S3 over AWS Direct Connect.
Answers
Suggested answer: B

An organization is developing a mobile social application and needs to collect logs from all devices on which it is installed. The organization is evaluating the Amazon Kinesis Data Streams to push logs and Amazon EMR to process data. They want to store data on HDFS using the default replication factor to replicate data among the cluster, but they are concerned about the durability of the data. Currently, they are producing 300 GB of raw data daily, with additional spikes during special events. They will need to scale out the Amazon EMR cluster to match the increase in streamed data. Which solution prevents data loss and matches compute demand?

A.
Use multiple Amazon EBS volumes on Amazon EMR to store processed data and scale out the Amazon EMR cluster as needed.
A.
Use multiple Amazon EBS volumes on Amazon EMR to store processed data and scale out the Amazon EMR cluster as needed.
Answers
B.
Use the EMR File System and Amazon S3 to store processed data and scale out the Amazon EMR cluster as needed.
B.
Use the EMR File System and Amazon S3 to store processed data and scale out the Amazon EMR cluster as needed.
Answers
C.
Use Amazon DynamoDB to store processed data and scale out the Amazon EMR cluster as needed.
C.
Use Amazon DynamoDB to store processed data and scale out the Amazon EMR cluster as needed.
Answers
D.
use Amazon Kinesis Data Firehose and, instead of using Amazon EMR, stream logs directly into Amazon Elasticsearch Service.
D.
use Amazon Kinesis Data Firehose and, instead of using Amazon EMR, stream logs directly into Amazon Elasticsearch Service.
Answers
Suggested answer: D

An advertising organization uses an application to process a stream of events that are received from clients in multiple unstructured formats.

The application does the following:

Transforms the events into a single structured format and streams them to Amazon Kinesis for real-time analysis.

Stores the unstructured raw events from the log files on local hard drivers that are rotated and uploaded to Amazon S3. The organization wants to extract campaign performance reporting using an existing Amazon redshift cluster.

Which solution will provide the performance data with the LEAST number of operations?

A.
Install the Amazon Kinesis Data Firehose agent on the application servers and use it to stream the log files directly to Amazon Redshift.
A.
Install the Amazon Kinesis Data Firehose agent on the application servers and use it to stream the log files directly to Amazon Redshift.
Answers
B.
Create an external table in Amazon Redshift and point it to the S3 bucket where the unstructured raw events are stored.
B.
Create an external table in Amazon Redshift and point it to the S3 bucket where the unstructured raw events are stored.
Answers
C.
Write an AWS Lambda function that triggers every hour to load the new log files already in S3 to Amazon redshift.
C.
Write an AWS Lambda function that triggers every hour to load the new log files already in S3 to Amazon redshift.
Answers
D.
Connect Amazon Kinesis Data Firehose to the existing Amazon Kinesis stream and use it to stream the event directly to Amazon Redshift.
D.
Connect Amazon Kinesis Data Firehose to the existing Amazon Kinesis stream and use it to stream the event directly to Amazon Redshift.
Answers
Suggested answer: B

An organization is designing an Amazon DynamoDB table for an application that must meet the following requirements: Item size is 40 KB

Read/write ratio 2000/500 sustained, respectively

Heavily read-oriented and requires low latencies in the order of milliseconds

The application runs on an Amazon EC2 instance

Access to the DynamoDB table must be secure within the VPC

Minimal changes to application code to improve performance using write-through cache Which design options will BEST meet these requirements?

A.
Size the DynamoDB table with 10000 RCUs/20000 WCUs, implement the DynamoDB Accelerator (DAX) for read performance, use VPC endpoints forDynamoDB, and implement an IAM role on the EC2 instance to secure DynamoDBaccess.
A.
Size the DynamoDB table with 10000 RCUs/20000 WCUs, implement the DynamoDB Accelerator (DAX) for read performance, use VPC endpoints forDynamoDB, and implement an IAM role on the EC2 instance to secure DynamoDBaccess.
Answers
B.
Size the DynamoDB table with 20000 RCUs/20000 WCUs, implement the DynamoDB Accelerator (DAX) for read performance, leverage VPC endpoints forDynamoDB, and implement an IAM user on the EC2 instance to secureDynamoDB access.
B.
Size the DynamoDB table with 20000 RCUs/20000 WCUs, implement the DynamoDB Accelerator (DAX) for read performance, leverage VPC endpoints forDynamoDB, and implement an IAM user on the EC2 instance to secureDynamoDB access.
Answers
C.
Size the DynamoDB table with 10000 RCUs/20000 WCUs, implement Amazon ElastiCache for read performance, set up a NAT gateway on VPC for the EC2instance to access DynamoDB, and implement an IAM role on the EC2instance to secure DynamoDB access.
C.
Size the DynamoDB table with 10000 RCUs/20000 WCUs, implement Amazon ElastiCache for read performance, set up a NAT gateway on VPC for the EC2instance to access DynamoDB, and implement an IAM role on the EC2instance to secure DynamoDB access.
Answers
D.
Size the DynamoDB table with 20000 RCUs/20000 WCUs, implement Amazon ElastiCache for read performance, leverage VPC endpoints for DynamoDB,and implement an IAM user on the EC2 instance to secure DynamoDB access.
D.
Size the DynamoDB table with 20000 RCUs/20000 WCUs, implement Amazon ElastiCache for read performance, leverage VPC endpoints for DynamoDB,and implement an IAM user on the EC2 instance to secure DynamoDB access.
Answers
Suggested answer: A

An organization needs to store sensitive information on Amazon S3 and process it through Amazon EMR. Data must be encrypted on Amazon S3 and Amazon EMR at rest and in transit. Using Thrift Server, the Data Analysis team users HIVE to interact with this data. The organization would like to grant access to only specific databases and tables, giving permission only to the SELECT statement.

Which solution will protect the data and limit user access to the SELECT statement on a specific portion of data?

A.
Configure Transparent Data Encryption on Amazon EMR. Create an Amazon EC2 instance and install Apache Ranger. Configure the authorization on thecluster to use Apache Ranger.
A.
Configure Transparent Data Encryption on Amazon EMR. Create an Amazon EC2 instance and install Apache Ranger. Configure the authorization on thecluster to use Apache Ranger.
Answers
B.
Configure data encryption at rest for EMR File System (EMRFS) on Amazon S3. Configure data encryption in transit for traffic between Amazon S3 andEMRFS. Configure storage and SQL base authorization on HiveServer2.
B.
Configure data encryption at rest for EMR File System (EMRFS) on Amazon S3. Configure data encryption in transit for traffic between Amazon S3 andEMRFS. Configure storage and SQL base authorization on HiveServer2.
Answers
C.
Use AWS KMS for encryption of data. Configure and attach multiple roles with different permissions based on the different user needs.
C.
Use AWS KMS for encryption of data. Configure and attach multiple roles with different permissions based on the different user needs.
Answers
D.
Configure Security Group on Amazon EMR. Create an Amazon VPC endpoint for Amazon S3. Configure HiveServer2 to use Kerberos authentication on thecluster.
D.
Configure Security Group on Amazon EMR. Create an Amazon VPC endpoint for Amazon S3. Configure HiveServer2 to use Kerberos authentication on thecluster.
Answers
Suggested answer: C

Multiple rows in an Amazon Redshift table were accidentally deleted. A System Administrator is restoring the table from the most recent snapshot. The snapshot contains all rows that were in the table before the deletion.

What is the SIMPLEST solution to restore the table without impacting users?

A.
Restore the snapshot to a new Amazon Redshift cluster, then UNLOAD the table to Amazon S3. In the original cluster, TRUNCATE the table, then load thedata from Amazon S3 by using a COPY command.
A.
Restore the snapshot to a new Amazon Redshift cluster, then UNLOAD the table to Amazon S3. In the original cluster, TRUNCATE the table, then load thedata from Amazon S3 by using a COPY command.
Answers
B.
Use the Restore Table from a Snapshot command and specify a new table name DROP the original table, then RENAME the new table to the original tablename.
B.
Use the Restore Table from a Snapshot command and specify a new table name DROP the original table, then RENAME the new table to the original tablename.
Answers
C.
Restore the snapshot to a new Amazon Redshift cluster. Create a DBLINK between the two clusters in the original cluster, TRUNCATE the destination table,then use an INSERT command to copy the data from the new cluster.
C.
Restore the snapshot to a new Amazon Redshift cluster. Create a DBLINK between the two clusters in the original cluster, TRUNCATE the destination table,then use an INSERT command to copy the data from the new cluster.
Answers
D.
Use the ALTER TABLE REVERT command and specify a time stamp of immediately before the data deletion. Specify the Amazon Resource Name of thesnapshot as the SOURCE and use the OVERWRITE REPLACE option.
D.
Use the ALTER TABLE REVERT command and specify a time stamp of immediately before the data deletion. Specify the Amazon Resource Name of thesnapshot as the SOURCE and use the OVERWRITE REPLACE option.
Answers
Suggested answer: B

An organization's data warehouse contains sales data for reporting purposes. data governance policies prohibit staff from accessing the customers' credit card numbers.

How can these policies be adhered to and still allow a Data Scientist to group transactions that use the same credit card number?

A.
Store a cryptographic hash of the credit card number.
A.
Store a cryptographic hash of the credit card number.
Answers
B.
Encrypt the credit card number with a symmetric encryption key, and give the key only to the authorized Data Scientist.
B.
Encrypt the credit card number with a symmetric encryption key, and give the key only to the authorized Data Scientist.
Answers
C.
Mask the credit card numbers to only show the last four digits of the credit card number.
C.
Mask the credit card numbers to only show the last four digits of the credit card number.
Answers
D.
Encrypt the credit card number with an asymmetric encryption key and give the decryption key only to the authorized Data Scientist.
D.
Encrypt the credit card number with an asymmetric encryption key and give the decryption key only to the authorized Data Scientist.
Answers
Suggested answer: C
Total 85 questions
Go to page: of 9