ExamGecko
Home Home / Amazon / DAS-C01

Amazon DAS-C01 Practice Test - Questions Answers, Page 16

Question list
Search
Search

List of questions

Search

Related questions











A data analyst is using Amazon QuickSight for data visualization across multiple datasets generated by applications. Each application stores files within a separate Amazon S3 bucket. AWS Glue Data Catalog is used as a central catalog across all application data in Amazon S3. A new application stores its data within a separate S3 bucket. After updating the catalog to include the new application data source, the data analyst created a new Amazon QuickSight data source from an Amazon Athena table, but the import into SPICE failed. How should the data analyst resolve the issue?

A.
Edit the permissions for the AWS Glue Data Catalog from within the Amazon QuickSight console.
A.
Edit the permissions for the AWS Glue Data Catalog from within the Amazon QuickSight console.
Answers
B.
Edit the permissions for the new S3 bucket from within the Amazon QuickSight console.
B.
Edit the permissions for the new S3 bucket from within the Amazon QuickSight console.
Answers
C.
Edit the permissions for the AWS Glue Data Catalog from within the AWS Glue console.
C.
Edit the permissions for the AWS Glue Data Catalog from within the AWS Glue console.
Answers
D.
Edit the permissions for the new S3 bucket from within the S3 console.
D.
Edit the permissions for the new S3 bucket from within the S3 console.
Answers
Suggested answer: B

Explanation:


Reference: https://aws.amazon.com/blogs/big-data/harmonize-query-and-visualize-data-from-various-providers-using-awsglue-amazon-athena-and-amazon-quicksight/

A company wants to use an automatic machine learning (ML) Random Cut Forest (RCF) algorithm to visualize complex realworld scenarios, such as detecting seasonality and trends, excluding outers, and imputing missing values. The team working on this project is non-technical and is looking for an out-of-the-box solution that will require the LEAST amount of management overhead. Which solution will meet these requirements?

A.
Use an AWS Glue ML transform to create a forecast and then use Amazon QuickSight to visualize the data.
A.
Use an AWS Glue ML transform to create a forecast and then use Amazon QuickSight to visualize the data.
Answers
B.
Use Amazon QuickSight to visualize the data and then use ML-powered forecasting to forecast the key business metrics.
B.
Use Amazon QuickSight to visualize the data and then use ML-powered forecasting to forecast the key business metrics.
Answers
C.
Use a pre-build ML AMI from the AWS Marketplace to create forecasts and then use Amazon QuickSight to visualize the data.
C.
Use a pre-build ML AMI from the AWS Marketplace to create forecasts and then use Amazon QuickSight to visualize the data.
Answers
D.
Use calculated fields to create a new forecast and then use Amazon QuickSight to visualize the data.
D.
Use calculated fields to create a new forecast and then use Amazon QuickSight to visualize the data.
Answers
Suggested answer: A

Explanation:


Reference: https://aws.amazon.com/blogs/big-data/query-visualize-and-forecast-trufactor-web-session-intelligence-with-awsdata-exchange/

A media company has been performing analytics on log data generated by its applications. There has been a recent increase in the number of concurrent analytics jobs running, and the overall performance of existing jobs is decreasing as the number of new jobs is increasing. The partitioned data is stored in Amazon S3 One Zone-Infrequent Access (S3 One Zone-IA) and the analytic processing is performed on Amazon EMR clusters using the EMR File System (EMRFS) with consistent view enabled. A data analyst has determined that it is taking longer for the EMR task nodes to list objects in Amazon S3. Which action would MOST likely increase the performance of accessing log data in Amazon S3?

A.
Use a hash function to create a random string and add that to the beginning of the object prefixes when storing the log data in Amazon S3.
A.
Use a hash function to create a random string and add that to the beginning of the object prefixes when storing the log data in Amazon S3.
Answers
B.
Use a lifecycle policy to change the S3 storage class to S3 Standard for the log data.
B.
Use a lifecycle policy to change the S3 storage class to S3 Standard for the log data.
Answers
C.
Increase the read capacity units (RCUs) for the shared Amazon DynamoDB table.
C.
Increase the read capacity units (RCUs) for the shared Amazon DynamoDB table.
Answers
D.
Redeploy the EMR clusters that are running slowly to a different Availability Zone.
D.
Redeploy the EMR clusters that are running slowly to a different Availability Zone.
Answers
Suggested answer: D

An airline has .csv-formatted data stored in Amazon S3 with an AWS Glue Data Catalog. Data analysts want to join this data with call center data stored in Amazon Redshift as part of a dally batch process. The Amazon Redshift cluster is already under a heavy load. The solution must be managed, serverless, well-functioning, and minimize the load on the existing Amazon Redshift cluster. The solution should also require minimal effort and development activity. Which solution meets these requirements?

A.
Unload the call center data from Amazon Redshift to Amazon S3 using an AWS Lambda function. Perform the join with AWS Glue ETL scripts.
A.
Unload the call center data from Amazon Redshift to Amazon S3 using an AWS Lambda function. Perform the join with AWS Glue ETL scripts.
Answers
B.
Export the call center data from Amazon Redshift using a Python shell in AWS Glue. Perform the join with AWS Glue ETL scripts.
B.
Export the call center data from Amazon Redshift using a Python shell in AWS Glue. Perform the join with AWS Glue ETL scripts.
Answers
C.
Create an external table using Amazon Redshift Spectrum for the call center data and perform the join with Amazon Redshift.
C.
Create an external table using Amazon Redshift Spectrum for the call center data and perform the join with Amazon Redshift.
Answers
D.
Export the call center data from Amazon Redshift to Amazon EMR using Apache Sqoop. Perform the join with Apache Hive.
D.
Export the call center data from Amazon Redshift to Amazon EMR using Apache Sqoop. Perform the join with Apache Hive.
Answers
Suggested answer: C

A social media company is using business intelligence tools to analyze its data for forecasting. The company is using Apache Kafka to ingest the low-velocity data in near-real time. The company wants to build dynamic dashboards with machine learning (ML) insights to forecast key business trends. The dashboards must provide hourly updates from data in Amazon S3. Various teams at the company want to view the dashboards by using Amazon QuickSight with ML insights. The solution also must correct the scalability problems that the company experiences when it uses its current architecture to ingest data. Which solution will MOST cost-effectively meet these requirements?

A.
Replace Kafka with Amazon Managed Streaming for Apache Kafka. Ingest the data by using AWS Lambda, and store the data in Amazon S3. Use QuickSight Standard edition to refresh the data in SPICE from Amazon S3 hourly andcreate a dynamic dashboard with forecasting and ML insights.
A.
Replace Kafka with Amazon Managed Streaming for Apache Kafka. Ingest the data by using AWS Lambda, and store the data in Amazon S3. Use QuickSight Standard edition to refresh the data in SPICE from Amazon S3 hourly andcreate a dynamic dashboard with forecasting and ML insights.
Answers
B.
Replace Kafka with an Amazon Kinesis data stream. Use an Amazon Kinesis Data Firehose delivery stream to consume the data and store the data in Amazon S3. Use QuickSight Enterprise edition to refresh the data in SPICE fromAmazon S3 hourly and create a dynamic dashboard with forecasting and ML insights.
B.
Replace Kafka with an Amazon Kinesis data stream. Use an Amazon Kinesis Data Firehose delivery stream to consume the data and store the data in Amazon S3. Use QuickSight Enterprise edition to refresh the data in SPICE fromAmazon S3 hourly and create a dynamic dashboard with forecasting and ML insights.
Answers
C.
Configure the Kafka-Kinesis-Connector to publish the data to an Amazon Kinesis Data Firehose delivery stream that is configured to store the data in Amazon S3. Use QuickSight Enterprise edition to refresh the data in SPICE fromAmazon S3 hourly and create a dynamic dashboard with forecasting and ML insights.
C.
Configure the Kafka-Kinesis-Connector to publish the data to an Amazon Kinesis Data Firehose delivery stream that is configured to store the data in Amazon S3. Use QuickSight Enterprise edition to refresh the data in SPICE fromAmazon S3 hourly and create a dynamic dashboard with forecasting and ML insights.
Answers
D.
Configure the Kafka-Kinesis-Connector to publish the data to an Amazon Kinesis Data Firehose delivery stream that is configured to store the data in Amazon S3. Configure an AWS Glue crawler to crawl the data. Use an AmazonAthena data source with QuickSight Standard edition to refresh the data in SPICE hourly and create a dynamic dashboard with forecasting and ML insights.
D.
Configure the Kafka-Kinesis-Connector to publish the data to an Amazon Kinesis Data Firehose delivery stream that is configured to store the data in Amazon S3. Configure an AWS Glue crawler to crawl the data. Use an AmazonAthena data source with QuickSight Standard edition to refresh the data in SPICE hourly and create a dynamic dashboard with forecasting and ML insights.
Answers
Suggested answer: B

Explanation:


Reference: https://noise.getoto.net/tag/amazon-kinesis-data-firehose/

A company owns facilities with IoT devices installed across the world. The company is using Amazon Kinesis Data Streams to stream data from the devices to Amazon S3. The company's operations team wants to get insights from the IoT data to monitor data quality at ingestion. The insights need to be derived in near-real time, and the output must be logged to Amazon DynamoDB for further analysis. Which solution meets these requirements?

A.
Connect Amazon Kinesis Data Analytics to analyze the stream data. Save the output to DynamoDB by using the default output from Kinesis Data Analytics.
A.
Connect Amazon Kinesis Data Analytics to analyze the stream data. Save the output to DynamoDB by using the default output from Kinesis Data Analytics.
Answers
B.
Connect Amazon Kinesis Data Analytics to analyze the stream data. Save the output to DynamoDB by using an AWS Lambda function.
B.
Connect Amazon Kinesis Data Analytics to analyze the stream data. Save the output to DynamoDB by using an AWS Lambda function.
Answers
C.
Connect Amazon Kinesis Data Firehose to analyze the stream data by using an AWS Lambda function. Save the output to DynamoDB by using the default output from Kinesis Data Firehose.
C.
Connect Amazon Kinesis Data Firehose to analyze the stream data by using an AWS Lambda function. Save the output to DynamoDB by using the default output from Kinesis Data Firehose.
Answers
D.
Connect Amazon Kinesis Data Firehose to analyze the stream data by using an AWS Lambda function. Save the data to Amazon S3. Then run an AWS Glue job on schedule to ingest the data into DynamoDB.
D.
Connect Amazon Kinesis Data Firehose to analyze the stream data by using an AWS Lambda function. Save the data to Amazon S3. Then run an AWS Glue job on schedule to ingest the data into DynamoDB.
Answers
Suggested answer: C

A company has a data lake on AWS that ingests sources of data from multiple business units and uses Amazon Athena for queries. The storage layer is Amazon S3 using the AWS Glue Data Catalog. The company wants to make the data available to its data scientists and business analysts. However, the company first needs to manage data access for Athena based on user roles and responsibilities. What should the company do to apply these access controls with the LEAST operational overhead?

A.
Define security policy-based rules for the users and applications by role in AWS Lake Formation.
A.
Define security policy-based rules for the users and applications by role in AWS Lake Formation.
Answers
B.
Define security policy-based rules for the users and applications by role in AWS Identity and Access Management (IAM).
B.
Define security policy-based rules for the users and applications by role in AWS Identity and Access Management (IAM).
Answers
C.
Define security policy-based rules for the tables and columns by role in AWS Glue.
C.
Define security policy-based rules for the tables and columns by role in AWS Glue.
Answers
D.
Define security policy-based rules for the tables and columns by role in AWS Identity and Access Management (IAM).
D.
Define security policy-based rules for the tables and columns by role in AWS Identity and Access Management (IAM).
Answers
Suggested answer: D

A media company wants to perform machine learning and analytics on the data residing in its Amazon S3 data lake. There are two data transformation requirements that will enable the consumers within the company to create reports: Daily transformations of 300 GB of data with different file formats landing in Amazon S3 at a scheduled time. One-time transformations of terabytes of archived data residing in the S3 data lake.

Which combination of solutions cost-effectively meets the company’s requirements for transforming the data? (Choose three.)

A.
For daily incoming data, use AWS Glue crawlers to scan and identify the schema.
A.
For daily incoming data, use AWS Glue crawlers to scan and identify the schema.
Answers
B.
For daily incoming data, use Amazon Athena to scan and identify the schema.
B.
For daily incoming data, use Amazon Athena to scan and identify the schema.
Answers
C.
For daily incoming data, use Amazon Redshift to perform transformations.
C.
For daily incoming data, use Amazon Redshift to perform transformations.
Answers
D.
For daily incoming data, use AWS Glue workflows with AWS Glue jobs to perform transformations.
D.
For daily incoming data, use AWS Glue workflows with AWS Glue jobs to perform transformations.
Answers
E.
For archived data, use Amazon EMR to perform data transformations.
E.
For archived data, use Amazon EMR to perform data transformations.
Answers
F.
For archived data, use Amazon SageMaker to perform data transformations.
F.
For archived data, use Amazon SageMaker to perform data transformations.
Answers
Suggested answer: B, C, D

A company currently uses Amazon Athena to query its global datasets. The regional data is stored in Amazon S3 in the useast- 1 and us-west-2 Regions. The data is not encrypted. To simplify the query process and manage it centrally, the company wants to use Athena in us-west-2 to query data from Amazon S3 in both Regions. The solution should be as lowcost as possible. What should the company do to achieve this goal?

A.
Use AWS DMS to migrate the AWS Glue Data Catalog from us-east-1 to us-west-2. Run Athena queries in us-west-2.
A.
Use AWS DMS to migrate the AWS Glue Data Catalog from us-east-1 to us-west-2. Run Athena queries in us-west-2.
Answers
B.
Run the AWS Glue crawler in us-west-2 to catalog datasets in all Regions. Once the data is crawled, run Athena queries in us-west-2.
B.
Run the AWS Glue crawler in us-west-2 to catalog datasets in all Regions. Once the data is crawled, run Athena queries in us-west-2.
Answers
C.
Enable cross-Region replication for the S3 buckets in us-east-1 to replicate data in us-west-2. Once the data is replicated in us-west-2, run the AWS Glue crawler there to update the AWS Glue Data Catalog in us-west-2 and run Athenaqueries.
C.
Enable cross-Region replication for the S3 buckets in us-east-1 to replicate data in us-west-2. Once the data is replicated in us-west-2, run the AWS Glue crawler there to update the AWS Glue Data Catalog in us-west-2 and run Athenaqueries.
Answers
D.
Update AWS Glue resource policies to provide us-east-1 AWS Glue Data Catalog access to us-west-2. Once the catalog in us-west-2 has access to the catalog in us-east-1, run Athena queries in uswest-2.
D.
Update AWS Glue resource policies to provide us-east-1 AWS Glue Data Catalog access to us-west-2. Once the catalog in us-west-2 has access to the catalog in us-east-1, run Athena queries in uswest-2.
Answers
Suggested answer: C

A company uses Amazon Redshift for its data warehousing needs. ETL jobs run every night to load data, apply business rules, and create aggregate tables for reporting. The company's data analysis, data science, and business intelligence teams use the data warehouse during regular business hours. The workload management is set to auto, and separate queues exist for each team with the priority set to NORMAL.

Recently, a sudden spike of read queries from the data analysis team has occurred at least twice daily, and queries wait in line for cluster resources. The company needs a solution that enables the data analysis team to avoid query queuing without impacting latency and the query times of other teams. Which solution meets these requirements?

A.
Increase the query priority to HIGHEST for the data analysis queue.
A.
Increase the query priority to HIGHEST for the data analysis queue.
Answers
B.
Configure the data analysis queue to enable concurrency scaling.
B.
Configure the data analysis queue to enable concurrency scaling.
Answers
C.
Create a query monitoring rule to add more cluster capacity for the data analysis queue when queries are waiting for resources.
C.
Create a query monitoring rule to add more cluster capacity for the data analysis queue when queries are waiting for resources.
Answers
D.
Use workload management query queue hopping to route the query to the next matching queue.
D.
Use workload management query queue hopping to route the query to the next matching queue.
Answers
Suggested answer: D
Total 214 questions
Go to page: of 22