Amazon DEA-C01 Practice Test - Questions Answers, Page 2
List of questions
Related questions
A company stores petabytes of data in thousands of Amazon S3 buckets in the S3 Standard storage class. The data supports analytics workloads that have unpredictable and variable data access patterns.
The company does not access some data for months. However, the company must be able to retrieve all data within milliseconds. The company needs to optimize S3 storage costs.
Which solution will meet these requirements with the LEAST operational overhead?
Use S3 Storage Lens standard metrics to determine when to move objects to more cost-optimized storage classes. Create S3 Lifecycle policies for the S3 buckets to move objects to cost-optimized storage classes. Continue to refine the S3 Lifecycle policies in the future to optimize storage costs.
Use S3 Storage Lens activity metrics to identify S3 buckets that the company accesses infrequently. Configure S3 Lifecycle rules to move objects from S3 Standard to the S3 Standard-Infrequent Access (S3 Standard-IA) and S3 Glacier storage classes based on the age of the data.
Use S3 Intelligent-Tiering. Activate the Deep Archive Access tier.
Use S3 Intelligent-Tiering. Use the default access tier.
During a security review, a company identified a vulnerability in an AWS Glue job. The company discovered that credentials to access an Amazon Redshift cluster were hard coded in the job script.
A data engineer must remediate the security vulnerability in the AWS Glue job. The solution must securely store the credentials.
Which combination of steps should the data engineer take to meet these requirements? (Choose two.)
Store the credentials in the AWS Glue job parameters.
Store the credentials in a configuration file that is in an Amazon S3 bucket.
Access the credentials from a configuration file that is in an Amazon S3 bucket by using the AWS Glue job.
Store the credentials in AWS Secrets Manager.
Grant the AWS Glue job 1AM role access to the stored credentials.
A data engineer is configuring an AWS Glue job to read data from an Amazon S3 bucket. The data engineer has set up the necessary AWS Glue connection details and an associated IAM role. However, when the data engineer attempts to run the AWS Glue job, the data engineer receives an error message that indicates that there are problems with the Amazon S3 VPC gateway endpoint.
The data engineer must resolve the error and connect the AWS Glue job to the S3 bucket.
Which solution will meet this requirement?
Update the AWS Glue security group to allow inbound traffic from the Amazon S3 VPC gateway endpoint.
Configure an S3 bucket policy to explicitly grant the AWS Glue job permissions to access the S3 bucket.
Review the AWS Glue job code to ensure that the AWS Glue connection details include a fully qualified domain name.
Verify that the VPC's route table includes inbound and outbound routes for the Amazon S3 VPC gateway endpoint.
A retail company has a customer data hub in an Amazon S3 bucket. Employees from many countries use the data hub to support company-wide analytics. A governance team must ensure that the company's data analysts can access data only for customers who are within the same country as the analysts.
Which solution will meet these requirements with the LEAST operational effort?
Create a separate table for each country's customer data. Provide access to each analyst based on the country that the analyst serves.
Register the S3 bucket as a data lake location in AWS Lake Formation. Use the Lake Formation row-level security features to enforce the company's access policies.
Move the data to AWS Regions that are close to the countries where the customers are. Provide access to each analyst based on the country that the analyst serves.
Load the data into Amazon Redshift. Create a view for each country. Create separate 1AM roles for each country to provide access to data from each country. Assign the appropriate roles to the analysts.
A media company wants to improve a system that recommends media content to customer based on user behavior and preferences. To improve the recommendation system, the company needs to incorporate insights from third-party datasets into the company's existing analytics platform.
The company wants to minimize the effort and time required to incorporate third-party datasets.
Which solution will meet these requirements with the LEAST operational overhead?
Use API calls to access and integrate third-party datasets from AWS Data Exchange.
Use API calls to access and integrate third-party datasets from AWS
Use Amazon Kinesis Data Streams to access and integrate third-party datasets from AWS CodeCommit repositories.
Use Amazon Kinesis Data Streams to access and integrate third-party datasets from Amazon Elastic Container Registry (Amazon ECR).
A financial company wants to implement a data mesh. The data mesh must support centralized data governance, data analysis, and data access control. The company has decided to use AWS Glue for data catalogs and extract, transform, and load (ETL) operations.
Which combination of AWS services will implement a data mesh? (Choose two.)
Use Amazon Aurora for data storage. Use an Amazon Redshift provisioned cluster for data analysis.
Use Amazon S3 for data storage. Use Amazon Athena for data analysis.
Use AWS Glue DataBrewfor centralized data governance and access control.
Use Amazon RDS for data storage. Use Amazon EMR for data analysis.
Use AWS Lake Formation for centralized data governance and access control.
A data engineer maintains custom Python scripts that perform a data formatting process that many AWS Lambda functions use. When the data engineer needs to modify the Python scripts, the data engineer must manually update all the Lambda functions.
The data engineer requires a less manual way to update the Lambda functions.
Which solution will meet this requirement?
Store a pointer to the custom Python scripts in the execution context object in a shared Amazon S3 bucket.
Package the custom Python scripts into Lambda layers. Apply the Lambda layers to the Lambda functions.
Store a pointer to the custom Python scripts in environment variables in a shared Amazon S3 bucket.
Assign the same alias to each Lambda function. Call reach Lambda function by specifying the function's alias.
A company created an extract, transform, and load (ETL) data pipeline in AWS Glue. A data engineer must crawl a table that is in Microsoft SQL Server. The data engineer needs to extract, transform, and load the output of the crawl to an Amazon S3 bucket. The data engineer also must orchestrate the data pipeline.
Which AWS service or feature will meet these requirements MOST cost-effectively?
AWS Step Functions
AWS Glue workflows
AWS Glue Studio
Amazon Managed Workflows for Apache Airflow (Amazon MWAA)
A financial services company stores financial data in Amazon Redshift. A data engineer wants to run real-time queries on the financial data to support a web-based trading application. The data engineer wants to run the queries from within the trading application.
Which solution will meet these requirements with the LEAST operational overhead?
Establish WebSocket connections to Amazon Redshift.
Use the Amazon Redshift Data API.
Set up Java Database Connectivity (JDBC) connections to Amazon Redshift.
Store frequently accessed data in Amazon S3. Use Amazon S3 Select to run the queries.
A company uses Amazon Athena for one-time queries against data that is in Amazon S3. The company has several use cases. The company must implement permission controls to separate query processes and access to query history among users, teams, and applications that are in the same AWS account.
Which solution will meet these requirements?
Create an S3 bucket for each use case. Create an S3 bucket policy that grants permissions to appropriate individual IAM users. Apply the S3 bucket policy to the S3 bucket.
Create an Athena workgroup for each use case. Apply tags to the workgroup. Create an 1AM policy that uses the tags to apply appropriate permissions to the workgroup.
Create an JAM role for each use case. Assign appropriate permissions to the role for each use case. Associate the role with Athena.
Create an AWS Glue Data Catalog resource policy that grants permissions to appropriate individual IAM users for each use case. Apply the resource policy to the specific tables that Athena uses.
Question