ExamGecko
Home Home / Amazon / DAS-C01

Amazon DAS-C01 Practice Test - Questions Answers, Page 6

Question list
Search
Search

List of questions

Search

Related questions











A reseller that has thousands of AWS accounts receives AWS Cost and Usage Reports in an Amazon S3 bucket. The reports are delivered to the S3 bucket in the following format:

//yyyymmdd-yyyymmdd/.parquet

An AWS Glue crawler crawls the S3 bucket and populates an AWS Glue Data Catalog with a table. Business analysts use Amazon Athena to query the table and create monthly summary reports for the AWS accounts. The business analysts are experiencing slow queries because of the accumulation of reports from the last 5 years. The business analysts want the operations team to make changes to improve query performance. Which action should the operations team take to meet these requirements?

A.
Change the file format to .csv.zip
A.
Change the file format to .csv.zip
Answers
B.
Partition the data by date and account ID
B.
Partition the data by date and account ID
Answers
C.
Partition the data by month and account ID
C.
Partition the data by month and account ID
Answers
D.
Partition the data by account ID, year, and month
D.
Partition the data by account ID, year, and month
Answers
Suggested answer: A

Explanation:


Reference: https://docs.aws.amazon.com/cur/latest/userguide/access-cur-s3.html

A company uses Amazon Redshift to store its data. The reporting team runs ad-hoc queries to generate reports from the Amazon Redshift database. The reporting team recently started to experience inconsistencies in report generation. Adhoc queries used to generate reports that would typically take minutes to run can take hours to run. A data analytics specialist debugging the issue finds that ad-hoc queries are stuck in the queue behind long-running queries. How should the data analytics specialist resolve the issue?

A.
Create partitions in the tables queried in ad-hoc queries.
A.
Create partitions in the tables queried in ad-hoc queries.
Answers
B.
Configure automatic workload management (WLM) from the Amazon Redshift console.
B.
Configure automatic workload management (WLM) from the Amazon Redshift console.
Answers
C.
Create Amazon Simple Queue Service (Amazon SQS) queues with different priorities. Assign queries to a queue based on priority.
C.
Create Amazon Simple Queue Service (Amazon SQS) queues with different priorities. Assign queries to a queue based on priority.
Answers
D.
Run the VACUUM command for all tables in the database.
D.
Run the VACUUM command for all tables in the database.
Answers
Suggested answer: C

Explanation:


Reference: https://aws.amazon.com/sqs/features/

A mobile gaming company wants to capture data from its gaming app and make the data available for analysis immediately.

The data record size will be approximately 20 KB. The company is concerned about achieving optimal throughput from each device. Additionally, the company wants to develop a data stream processing application with dedicated throughput for each consumer.

Which solution would achieve this goal?

A.
Have the app call the PutRecords API to send data to Amazon Kinesis Data Streams. Use the enhanced fan-out feature while consuming the data.
A.
Have the app call the PutRecords API to send data to Amazon Kinesis Data Streams. Use the enhanced fan-out feature while consuming the data.
Answers
B.
Have the app call the PutRecordBatch API to send data to Amazon Kinesis Data Firehose. Submit a support case to enable dedicated throughput on the account.
B.
Have the app call the PutRecordBatch API to send data to Amazon Kinesis Data Firehose. Submit a support case to enable dedicated throughput on the account.
Answers
C.
Have the app use Amazon Kinesis Producer Library (KPL) to send data to Kinesis Data Firehose. Use the enhanced fanout feature while consuming the data.
C.
Have the app use Amazon Kinesis Producer Library (KPL) to send data to Kinesis Data Firehose. Use the enhanced fanout feature while consuming the data.
Answers
D.
Have the app call the PutRecords API to send data to Amazon Kinesis Data Streams. Host the stream-processing application on Amazon EC2 with Auto Scaling.
D.
Have the app call the PutRecords API to send data to Amazon Kinesis Data Streams. Host the stream-processing application on Amazon EC2 with Auto Scaling.
Answers
Suggested answer: D

A large company has a central data lake to run analytics across different departments. Each department uses a separate AWS account and stores its data in an Amazon S3 bucket in that account. Each AWS account uses the AWS Glue Data Catalog as its data catalog. There are different data lake access requirements based on roles. Associate analysts should only have read access to their departmental data. Senior data analysts can have access in multiple departments including theirs, but for a subset of columns only.

Which solution achieves these required access patterns to minimize costs and administrative tasks?

A.
Consolidate all AWS accounts into one account. Create different S3 buckets for each department and move all the data from every account to the central data lake account. Migrate the individual data catalogs into a central data catalogand apply fine-grained permissions to give to each user the required access to tables and databases in AWS Glue and Amazon S3.
A.
Consolidate all AWS accounts into one account. Create different S3 buckets for each department and move all the data from every account to the central data lake account. Migrate the individual data catalogs into a central data catalogand apply fine-grained permissions to give to each user the required access to tables and databases in AWS Glue and Amazon S3.
Answers
B.
Keep the account structure and the individual AWS Glue catalogs on each account. Add a central data lake account and use AWS Glue to catalog data from various accounts. Configure cross-account access for AWS Glue crawlers toscan the data in each departmental S3 bucket to identify the schema and populate the catalog. Add the senior data analysts into the central account and apply highly detailed access controls in the Data Catalog and Amazon S3.
B.
Keep the account structure and the individual AWS Glue catalogs on each account. Add a central data lake account and use AWS Glue to catalog data from various accounts. Configure cross-account access for AWS Glue crawlers toscan the data in each departmental S3 bucket to identify the schema and populate the catalog. Add the senior data analysts into the central account and apply highly detailed access controls in the Data Catalog and Amazon S3.
Answers
C.
Set up an individual AWS account for the central data lake. Use AWS Lake Formation to catalog the cross-account locations. On each individual S3 bucket, modify the bucket policy to grant S3 permissions to the Lake Formationservicelinked role. Use Lake Formation permissions to add fine-grained access controls to allow senior analysts to view specific tables and columns.
C.
Set up an individual AWS account for the central data lake. Use AWS Lake Formation to catalog the cross-account locations. On each individual S3 bucket, modify the bucket policy to grant S3 permissions to the Lake Formationservicelinked role. Use Lake Formation permissions to add fine-grained access controls to allow senior analysts to view specific tables and columns.
Answers
D.
Set up an individual AWS account for the central data lake and configure a central S3 bucket. Use an AWS Lake Formation blueprint to move the data from the various buckets into the central S3 bucket. On each individual bucket,modify the bucket policy to grant S3 permissions to the Lake Formation service-linked role. Use Lake Formation permissions to add fine-grained access controls for both associate and senior analysts to view specific tables and columns.
D.
Set up an individual AWS account for the central data lake and configure a central S3 bucket. Use an AWS Lake Formation blueprint to move the data from the various buckets into the central S3 bucket. On each individual bucket,modify the bucket policy to grant S3 permissions to the Lake Formation service-linked role. Use Lake Formation permissions to add fine-grained access controls for both associate and senior analysts to view specific tables and columns.
Answers
Suggested answer: B

A manufacturing company is storing data from its operational systems in Amazon S3. The company’s business analysts need to perform one-time queries of the data in Amazon S3 with Amazon Athena. The company needs to access the Athena network from the on-premises network by using a JDBC connection. The company has created a VPC Security policies mandate that requests to AWS services cannot traverse the Internet.

Which combination of steps should a data analytics specialist take to meet these requirements? (Choose two.)

A.
Establish an AWS Direct Connect connection between the on-premises network and the VPC.
A.
Establish an AWS Direct Connect connection between the on-premises network and the VPC.
Answers
B.
Configure the JDBC connection to connect to Athena through Amazon API Gateway.
B.
Configure the JDBC connection to connect to Athena through Amazon API Gateway.
Answers
C.
Configure the JDBC connection to use a gateway VPC endpoint for Amazon S3.
C.
Configure the JDBC connection to use a gateway VPC endpoint for Amazon S3.
Answers
D.
Configure the JDBC connection to use an interface VPC endpoint for Athena.
D.
Configure the JDBC connection to use an interface VPC endpoint for Athena.
Answers
E.
Deploy Athena within a private subnet.
E.
Deploy Athena within a private subnet.
Answers
Suggested answer: A, E

Explanation:


AWS Direct Connect makes it easy to establish a dedicated connection from an on-premises network to one or more VPCs in the same region. Reference: https://docs.aws.amazon.com/whitepapers/latest/aws-vpc-connectivity-options/aws-direct-connect.html https://stackoverflow.com/questions/68798311/aws-athena-connect-from-lambda

A company receives data from its vendor in JSON format with a timestamp in the file name. The vendor uploads the data to an Amazon S3 bucket, and the data is registered into the company’s data lake for analysis and reporting. The company has configured an S3 Lifecycle policy to archive all files to S3 Glacier after 5 days.

The company wants to ensure that its AWS Glue crawler catalogs data only from S3 Standard storage and ignores the archived files. A data analytics specialist must implement a solution to achieve this goal without changing the current S3 bucket configuration.

Which solution meets these requirements?

A.
Use the exclude patterns feature of AWS Glue to identify the S3 Glacier files for the crawler to exclude.
A.
Use the exclude patterns feature of AWS Glue to identify the S3 Glacier files for the crawler to exclude.
Answers
B.
Schedule an automation job that uses AWS Lambda to move files from the original S3 bucket to a new S3 bucket for S3 Glacier storage.
B.
Schedule an automation job that uses AWS Lambda to move files from the original S3 bucket to a new S3 bucket for S3 Glacier storage.
Answers
C.
Use the excludeStorageClasses property in the AWS Glue Data Catalog table to exclude files on S3 Glacier storage.
C.
Use the excludeStorageClasses property in the AWS Glue Data Catalog table to exclude files on S3 Glacier storage.
Answers
D.
Use the include patterns feature of AWS Glue to identify the S3 Standard files for the crawler to include.
D.
Use the include patterns feature of AWS Glue to identify the S3 Standard files for the crawler to include.
Answers
Suggested answer: A

Explanation:


Reference: https://docs.aws.amazon.com/glue/latest/dg/define-crawler.html#crawler-data-stores-exclude

A US-based sneaker retail company launched its global website. All the transaction data is stored in Amazon RDS and curated historic transaction data is stored in Amazon Redshift in the us-east-1 Region. The business intelligence (BI) team wants to enhance the user experience by providing a dashboard for sneaker trends.

The BI team decides to use Amazon QuickSight to render the website dashboards. During development, a team in Japan provisioned Amazon QuickSight in ap-northeast-1. The team is having difficulty connecting Amazon QuickSight from apnortheast- 1 to Amazon Redshift in us-east-1.

Which solution will solve this issue and meet the requirements?

A.
In the Amazon Redshift console, choose to configure cross-Region snapshots and set the destination Region as apnortheast- 1. Restore the Amazon Redshift Cluster from the snapshot and connect to Amazon QuickSight launched inapnortheast- 1.
A.
In the Amazon Redshift console, choose to configure cross-Region snapshots and set the destination Region as apnortheast- 1. Restore the Amazon Redshift Cluster from the snapshot and connect to Amazon QuickSight launched inapnortheast- 1.
Answers
B.
Create a VPC endpoint from the Amazon QuickSight VPC to the Amazon Redshift VPC so Amazon QuickSight can access data from Amazon Redshift.
B.
Create a VPC endpoint from the Amazon QuickSight VPC to the Amazon Redshift VPC so Amazon QuickSight can access data from Amazon Redshift.
Answers
C.
Create an Amazon Redshift endpoint connection string with Region information in the string and use this connection string in Amazon QuickSight to connect to Amazon Redshift.
C.
Create an Amazon Redshift endpoint connection string with Region information in the string and use this connection string in Amazon QuickSight to connect to Amazon Redshift.
Answers
D.
Create a new security group for Amazon Redshift in us-east-1 with an inbound rule authorizing access from the appropriate IP address range for the Amazon QuickSight servers in ap-northeast-1.
D.
Create a new security group for Amazon Redshift in us-east-1 with an inbound rule authorizing access from the appropriate IP address range for the Amazon QuickSight servers in ap-northeast-1.
Answers
Suggested answer: B

A company has developed several AWS Glue jobs to validate and transform its data from Amazon S3 and load it into Amazon RDS for MySQL in batches once every day. The ETL jobs read the S3 data using a DynamicFrame. Currently, the ETL developers are experiencing challenges in processing only the incremental data on every run, as the AWS Glue job processes all the S3 input data on each run.

Which approach would allow the developers to solve the issue with minimal coding effort?

A.
Have the ETL jobs read the data from Amazon S3 using a DataFrame.
A.
Have the ETL jobs read the data from Amazon S3 using a DataFrame.
Answers
B.
Enable job bookmarks on the AWS Glue jobs.
B.
Enable job bookmarks on the AWS Glue jobs.
Answers
C.
Create custom logic on the ETL jobs to track the processed S3 objects.
C.
Create custom logic on the ETL jobs to track the processed S3 objects.
Answers
D.
Have the ETL jobs delete the processed objects or data from Amazon S3 after each run.
D.
Have the ETL jobs delete the processed objects or data from Amazon S3 after each run.
Answers
Suggested answer: D

A manufacturing company wants to create an operational analytics dashboard to visualize metrics from equipment in nearreal time. The company uses Amazon Kinesis Data Streams to stream the data to other applications. The dashboard must automatically refresh every 5 seconds. A data analytics specialist must design a solution that requires the least possible implementation effort. Which solution meets these requirements?

A.
Use Amazon Kinesis Data Firehose to store the data in Amazon S3. Use Amazon QuickSight to build the dashboard.
A.
Use Amazon Kinesis Data Firehose to store the data in Amazon S3. Use Amazon QuickSight to build the dashboard.
Answers
B.
Use Apache Spark Streaming on Amazon EMR to read the data in near-real time. Develop a custom application for the dashboard by using D3.js.
B.
Use Apache Spark Streaming on Amazon EMR to read the data in near-real time. Develop a custom application for the dashboard by using D3.js.
Answers
C.
Use Amazon Kinesis Data Firehose to push the data into an Amazon OpenSearch Service (Amazon Elasticsearch Service) cluster. Visualize the data by using an OpenSearch Dashboards (Kibana).
C.
Use Amazon Kinesis Data Firehose to push the data into an Amazon OpenSearch Service (Amazon Elasticsearch Service) cluster. Visualize the data by using an OpenSearch Dashboards (Kibana).
Answers
D.
Use AWS Glue streaming ETL to store the data in Amazon S3. Use Amazon QuickSight to build the dashboard.
D.
Use AWS Glue streaming ETL to store the data in Amazon S3. Use Amazon QuickSight to build the dashboard.
Answers
Suggested answer: B

Explanation:


Reference: https://aws.amazon.com/blogs/big-data/analyze-a-time-series-in-real-time-with-aws-lambda-amazon-kinesis-andamazon-dynamodb-streams/

A real estate company maintains data about all properties listed in a market. The company receives data about new property listings from vendors who upload the data daily as compressed files into Amazon S3. The company’s leadership team wants to see the most up-to-date listings as soon as the data is uploaded to Amazon S3. The data analytics team must automate and orchestrate the data processing workflow of the listings to feed a dashboard. The team also must provide the ability to perform one-time queries and analytical reporting in a scalable manner. Which solution meets these requirements MOST cost-effectively?

A.
Use Amazon EMR for processing incoming data. Use AWS Step Functions for workflow orchestration. Use Apache Hive for one-time queries and analytical reporting. Bulk ingest the data in Amazon OpenSearch Service (AmazonElasticsearch Service). Use OpenSearch Dashboards (Kibana) on Amazon OpenSearch Service (Amazon Elasticsearch Service) for the dashboard.
A.
Use Amazon EMR for processing incoming data. Use AWS Step Functions for workflow orchestration. Use Apache Hive for one-time queries and analytical reporting. Bulk ingest the data in Amazon OpenSearch Service (AmazonElasticsearch Service). Use OpenSearch Dashboards (Kibana) on Amazon OpenSearch Service (Amazon Elasticsearch Service) for the dashboard.
Answers
B.
Use Amazon EMR for processing incoming data. Use AWS Step Functions for workflow orchestration. Use Amazon Athena for one-time queries and analytical reporting. Use Amazon QuickSight for the dashboard.
B.
Use Amazon EMR for processing incoming data. Use AWS Step Functions for workflow orchestration. Use Amazon Athena for one-time queries and analytical reporting. Use Amazon QuickSight for the dashboard.
Answers
C.
Use AWS Glue for processing incoming data. Use AWS Step Functions for workflow orchestration. Use Amazon Redshift Spectrum for one-time queries and analytical reporting. Use OpenSearch Dashboards (Kibana) on AmazonOpenSearch Service (Amazon Elasticsearch Service) for the dashboard.
C.
Use AWS Glue for processing incoming data. Use AWS Step Functions for workflow orchestration. Use Amazon Redshift Spectrum for one-time queries and analytical reporting. Use OpenSearch Dashboards (Kibana) on AmazonOpenSearch Service (Amazon Elasticsearch Service) for the dashboard.
Answers
D.
Use AWS Glue for processing incoming data. Use AWS Lambda and S3 Event Notifications for workflow orchestration.Use Amazon Athena for one-time queries and analytical reporting. Use Amazon QuickSight for the dashboard.
D.
Use AWS Glue for processing incoming data. Use AWS Lambda and S3 Event Notifications for workflow orchestration.Use Amazon Athena for one-time queries and analytical reporting. Use Amazon QuickSight for the dashboard.
Answers
Suggested answer: B

Explanation:


Reference: https://aws.amazon.com/blogs/compute/visualizing-aws-step-functions-workflows-from-the-amazon-athenaconsole/

Total 214 questions
Go to page: of 22