ExamGecko
Home Home / Microsoft / DP-203

Microsoft DP-203 Practice Test - Questions Answers, Page 21

Question list
Search
Search

List of questions

Search

Related questions











DRAG DROP

You have an Azure Data Lake Storage Gen2 account that contains a JSON file for customers. The file contains two attributes named FirstName and LastName. You need to copy the data from the JSON file to an Azure Synapse Analytics table by using Azure Databricks. A new column must be created that concatenates the FirstName and LastName values. You create the following components:

A destination table in Azure Synapse

An Azure Blob storage container

A service principal

In which order should you perform the actions? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.

Question 201
Correct answer: Question 201

Explanation:

Step 1: Mount the Data Lake Storage onto DBFS

Begin with creating a file system in the Azure Data Lake Storage Gen2 account. Step 2: Read the file into a data frame.

You can load the json files as a data frame in Azure Databricks. Step 3: Perform transformations on the data frame.

Step 4: Specify a temporary folder to stage the data

Specify a temporary folder to use while moving data between Azure Databricks and Azure Synapse. Step 5: Write the results to a table in Azure Synapse.

You upload the transformed data frame into Azure Synapse. You use the Azure Synapse connector for Azure Databricks to directly upload a dataframe as a table in a Azure Synapse.

Reference: https://docs.microsoft.com/en-us/azure/azure-databricks/databricks-extract-load-sql-data-warehouse

HOTSPOT

You need to design a data ingestion and storage solution for the Twitter feeds. The solution must meet the customer sentiment analytics requirements.

What should you include in the solution? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

Question 202
Correct answer: Question 202

Explanation:

Box 1: Configure Evegent Hubs partitions

Scenario: Maximize the throughput of ingesting Twitter feeds from Event Hubs to Azure Storage without purchasing additional throughput or capacity units.

Event Hubs is designed to help with processing of large volumes of events. Event Hubs throughput is scaled by using partitions and throughput-unit allocations.

Incorrect Answers:

Event Hubs Dedicated: Event Hubs clusters offer single-tenant deployments for customers with the most demanding streaming needs. This single-tenant offering has a guaranteed 99.99% SLA and is available only on our Dedicated pricing tier.

Auto-Inflate: The Auto-inflate feature of Event Hubs automatically scales up by increasing the number of TUs, to meet usage needs.

Event Hubs traffic is controlled by TUs (standard tier). Auto-inflate enables you to start small with the minimum required TUs you choose. The feature then scales automatically to the maximum limit of TUs you need, depending on the increase in your traffic.

Box 2: An Azure Data Lake Storage Gen2 account

Scenario: Ensure that the data store supports Azure AD-based access control down to the object level. Azure Data Lake Storage Gen2 implements an access control model that supports both Azure role-based access control (Azure RBAC) and POSIX-like access control lists (ACLs).

Incorrect Answers:

Azure Databricks: An Azure administrator with the proper permissions can configure Azure Active Directory conditional access to control where and when users are permitted to sign in to Azure Databricks. Azure Storage supports using Azure Active Directory (Azure AD) to authorize requests to blob data. You can scope access to Azure blob resources at the following levels, beginning with the narrowest scope:

- An individual container. At this scope, a role assignment applies to all of the blobs in the container, as well as container properties and metadata.

- The storage account. At this scope, a role assignment applies to all containers and their blobs. - The resource group. At this scope, a role assignment applies to all of the containers in all of the storage accounts in the resource group.

- The subscription. At this scope, a role assignment applies to all of the containers in all of the storage accounts in all of the resource groups in the subscription. - A management group.

Reference: https://docs.microsoft.com/en-us/azure/event-hubs/event-hubs-features https://docs.microsoft.com/en-us/azure/storage/blobs/data-lake-storage-access-control

You have an Azure Data Factory pipeline that is triggered hourly. The pipeline has had 100% success for the past seven days.

The pipeline execution fails, and two retries that occur 15 minutes apart also fail. The third failure returns the following error. ErrorCode=UserErrorFileNotFound,'Type=Microsoft.DataTransfer.Common.Shared.HybridDeliveryException,Message=ADL S Gen2 operation failed for: Operation returned an invalid status code 'NotFound'. Account: 'contosoproduksouth'. Filesystem: wwi. Path: 'BIKES/CARBON/year=2021/month=01/day=10/hour=06'. ErrorCode: 'PathNotFound'. Message: 'The specified path does not exist.'. RequestId: '6d269b78-901f-001b-4924-e7a7bc000000'. TimeStamp: 'Sun, 10 Jan 2021 07:45:05

What is a possible cause of the error?

A.
The parameter used to generate year=2021/month=01/day=10/hour=06 was incorrect.
A.
The parameter used to generate year=2021/month=01/day=10/hour=06 was incorrect.
Answers
B.
From 06:00 to 07:00 on January 10, 2021, there was no data in wwi/BIKES/CARBON.
B.
From 06:00 to 07:00 on January 10, 2021, there was no data in wwi/BIKES/CARBON.
Answers
C.
From 06:00 to 07:00 on January 10, 2021, the file format of data in wwi/BIKES/CARBON was incorrect.
C.
From 06:00 to 07:00 on January 10, 2021, the file format of data in wwi/BIKES/CARBON was incorrect.
Answers
D.
The pipeline was triggered too early.
D.
The pipeline was triggered too early.
Answers
Suggested answer: C

Explanation:

A file is missing.

You have an Azure Synapse Analytics job that uses Scala.

You need to view the status of the job.

What should you do?

A.
From Synapse Studio, select the workspace. From Monitor, select SQL requests.
A.
From Synapse Studio, select the workspace. From Monitor, select SQL requests.
Answers
B.
From Azure Monitor, run a Kusto query against the AzureDiagnostics table.
B.
From Azure Monitor, run a Kusto query against the AzureDiagnostics table.
Answers
C.
From Synapse Studio, select the workspace. From Monitor, select Apache Sparks applications.
C.
From Synapse Studio, select the workspace. From Monitor, select Apache Sparks applications.
Answers
D.
From Azure Monitor, run a Kusto query against the SparkLoggingEvent_CL table.
D.
From Azure Monitor, run a Kusto query against the SparkLoggingEvent_CL table.
Answers
Suggested answer: C

Explanation:

Use Synapse Studio to monitor your Apache Spark applications. To monitor running Apache Spark application Open Monitor, then select Apache Spark applications. To view the details about the Apache Spark applications that are running, select the submitting Apache Spark application and view the details. If the Apache Spark application is still running, you can monitor the progress.

Reference:

https://docs.microsoft.com/en-us/azure/synapse-analytics/monitoring/apache-spark-applications

You are designing database for an Azure Synapse Analytics dedicated SQL pool to support workloads for detecting ecommerce transaction fraud. Data will be combined from multiple ecommerce sites and can include sensitive financial information such as credit card numbers. You need to recommend a solution that meets the following requirements:

Users must be able to identify potentially fraudulent transactions. Users must be able to use credit cards as a potential feature in models. Users must NOT be able to access the actual credit card numbers. What should you include in the recommendation?

A.
Transparent Data Encryption (TDE)
A.
Transparent Data Encryption (TDE)
Answers
B.
row-level security (RLS)
B.
row-level security (RLS)
Answers
C.
column-level encryption
C.
column-level encryption
Answers
D.
Azure Active Directory (Azure AD) pass-through authentication
D.
Azure Active Directory (Azure AD) pass-through authentication
Answers
Suggested answer: B

Explanation:


You have an Azure subscription linked to an Azure Active Directory (Azure AD) tenant that contains a service principal named ServicePrincipal1. The subscription contains an Azure Data Lake Storage account named adls1. Adls1 contains a folder named Folder2 that has a URI of https://adls1.dfs.core.windows.net/container1/Folder1/Folder2/. ServicePrincipal1 has the access control list (ACL) permissions shown in the following table.

You need to ensure that ServicePrincipal1 can perform the following actions:

Traverse child items that are created in Folder2. Read files that are created in Folder2. The solution must use the principle of least privilege.

Which two permissions should you grant to ServicePrincipal1 for Folder2? Each correct answer presents part of the solution. NOTE: Each correct selection is worth one point.

A.
Access - Read
A.
Access - Read
Answers
B.
Access - Write
B.
Access - Write
Answers
C.
Access - Execute
C.
Access - Execute
Answers
D.
Default - Read
D.
Default - Read
Answers
E.
Default - Write
E.
Default - Write
Answers
F.
Default - Execute
F.
Default - Execute
Answers
Suggested answer: D, F

Explanation:

Execute (X) permission is required to traverse the child items of a folder. There are two kinds of access control lists (ACLs), Access ACLs and Default ACLs. Access ACLs: These control access to an object. Files and folders both have Access ACLs. Default ACLs: A "template" of ACLs associated with a folder that determine the Access ACLs for any child items that are created under that folder. Files do not have Default ACLs. Reference:

https://docs.microsoft.com/en-us/azure/data-lake-store/data-lake-store-access-control

You manage an enterprise data warehouse in Azure Synapse Analytics. Users report slow performance when they run commonly used queries. Users do not report performance changes for infrequently used queries. You need to monitor resource utilization to determine the source of the performance issues. Which metric should you monitor?

A.
Local tempdb percentage
A.
Local tempdb percentage
Answers
B.
Cache used percentage
B.
Cache used percentage
Answers
C.
Data IO percentage
C.
Data IO percentage
Answers
D.
CPU percentage
D.
CPU percentage
Answers
Suggested answer: B

Explanation:

Monitor and troubleshoot slow query performance by determining whether your workload is optimally leveraging the adaptive cache for dedicated SQL pools. Reference:

https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/sql-data-warehouse-how-to-monitor-cache

You have an Azure data factory.

You need to examine the pipeline failures from the last 180 days. What should you use?

A.
the Activity log blade for the Data Factory resource
A.
the Activity log blade for the Data Factory resource
Answers
B.
Pipeline runs in the Azure Data Factory user experience
B.
Pipeline runs in the Azure Data Factory user experience
Answers
C.
the Resource health blade for the Data Factory resource
C.
the Resource health blade for the Data Factory resource
Answers
D.
Azure Data Factory activity runs in Azure Monitor
D.
Azure Data Factory activity runs in Azure Monitor
Answers
Suggested answer: D

Explanation:

Data Factory stores pipeline-run data for only 45 days. Use Azure Monitor if you want to keep that data for a longer time.

Reference:

https://docs.microsoft.com/en-us/azure/data-factory/monitor-using-azure-monitor

HOTSPOT

You plan to create an Azure Data Lake Storage Gen2 account. You need to recommend a storage solution that meets the following requirements:

Provides the highest degree of data resiliency

Ensures that content remains available for writes if a primary data center fails What should you include in the recommendation? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

Question 209
Correct answer: Question 209

Explanation:

https://docs.microsoft.com/en-us/azure/storage/common/storage-disaster-recovery-guidance?toc=/azure/storage/blobs/toc.json https://docs.microsoft.com/en-us/answers/questions/32583/azure-data-lakegen2-disaster-recoverystorage-acco.html

You have an Azure data factory named ADF1.

You currently publish all pipeline authoring changes directly to ADF1. You need to implement version control for the changes made to pipeline artifacts. The solution must ensure that you can apply version control to the resources currently defined in the UX Authoring canvas for ADF1. Which two actions should you perform? Each correct answer presents part of the solution. NOTE: Each correct selection is worth one point.

A.
From the UX Authoring canvas, select Set up code repository.
A.
From the UX Authoring canvas, select Set up code repository.
Answers
B.
Create a Git repository.
B.
Create a Git repository.
Answers
C.
Create a GitHub action.
C.
Create a GitHub action.
Answers
D.
Create an Azure Data Factory trigger.
D.
Create an Azure Data Factory trigger.
Answers
E.
From the UX Authoring canvas, select Publish.
E.
From the UX Authoring canvas, select Publish.
Answers
F.
From the UX Authoring canvas, run Publish All.
F.
From the UX Authoring canvas, run Publish All.
Answers
Suggested answer: B, F

Explanation:

https://docs.microsoft.com/en-us/azure/data-factory/source-control

Total 320 questions
Go to page: of 32