ExamGecko
Home Home / Microsoft / DP-203

Microsoft DP-203 Practice Test - Questions Answers, Page 23

Question list
Search
Search

List of questions

Search

Related questions











DRAG DROP

You have an Azure subscription.

You plan to build a data warehouse in an Azure Synapse Analytics dedicated SQL pool named pool1 that will contain staging tables and a dimensional model Pool1 will contain the following tables.

Question 221
Correct answer: Question 221

You plan to create a dimension table in Azure Synapse Analytics that will be less than 1 GB. You need to create the table to meet the following requirements:

• Provide the fastest Query time.

• Minimize data movement during queries.

Which type of table should you use?

A.
hash distributed
A.
hash distributed
Answers
B.
heap
B.
heap
Answers
C.
replicated
C.
replicated
Answers
D.
round-robin
D.
round-robin
Answers
Suggested answer: C

Explanation:

A replicated table has a full copy of the table accessible on each Compute node. Replicating a table removes the need to transfer data among Compute nodes before a join or aggregation. Since the table has multiple copies, replicated tables work best when the table size is less than 2 GB compressed. 2 GB is not a hard limit.

Reference: https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/designguidance-for-replicated-tables

You haw an Azure data factory named ADF1.

You currently publish all pipeline authoring changes directly to ADF1. You need to implement version control for the changes made to pipeline artifacts. The solution must ensure that you can apply version control to the resources currently defined m the UX Authoring canvas for ADF1.

Which two actions should you perform? Each correct answer presents part of the solution NOTE: Each correct selection is worth one point.

A.
Create an Azure Data Factory trigger
A.
Create an Azure Data Factory trigger
Answers
B.
From the UX Authoring canvas, select Set up code repository
B.
From the UX Authoring canvas, select Set up code repository
Answers
C.
Create a GitHub action
C.
Create a GitHub action
Answers
D.
From the UX Authoring canvas, run Publish All.
D.
From the UX Authoring canvas, run Publish All.
Answers
E.
Create a Git repository
E.
Create a Git repository
Answers
F.
From the UX Authoring canvas, select Publish
F.
From the UX Authoring canvas, select Publish
Answers
Suggested answer: B, E

Explanation:

https://docs.microsoft.com/en-us/azure/data-factory/source-control

You haw an Azure data factory named ADF1.

You currently publish all pipeline authoring changes directly to ADF1. You need to implement version control for the changes made to pipeline artifacts. The solution must ensure that you can apply version control to the resources currently defined m the UX Authoring canvas for ADF1. Which two actions should you perform? Each correct answer presents part of the solution NOTE: Each correct selection is worth one point.

A.
Create an Azure Data Factory trigger
A.
Create an Azure Data Factory trigger
Answers
B.
From the UX Authoring canvas, select Set up code repository
B.
From the UX Authoring canvas, select Set up code repository
Answers
C.
Create a GitHub action
C.
Create a GitHub action
Answers
D.
From the UX Authoring canvas, run Publish All.
D.
From the UX Authoring canvas, run Publish All.
Answers
E.
Create a Git repository
E.
Create a Git repository
Answers
F.
From the UX Authoring canvas, select Publish
F.
From the UX Authoring canvas, select Publish
Answers
Suggested answer: B, D

You have an Azure Synapse Analytics dedicated SQL pool.

You need to Create a fact table named Table1 that will store sales data from the last three years. The solution must be optimized for the following query operations:

Show order counts by week.

• Calculate sales totals by region.

• Calculate sales totals by product.

• Find all the orders from a given month.

Which data should you use to partition Table1?

A.
region
A.
region
Answers
B.
product
B.
product
Answers
C.
week
C.
week
Answers
D.
month
D.
month
Answers
Suggested answer: C

HOTSPOT

You have an Azure Synapse Analytics serverless SQL pool, an Azure Synapse Analytics dedicated SQL pool, an Apache Spark pool, and an Azure Data Lake Storage Gen2 account. You need to create a table in a lake database. The table must be available to both the serverless SQL pool and the Spark pool. Where should you create the table, and Which file format should you use for data in the table? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

Question 226
Correct answer: Question 226

HOTSPOT

You have an Azure subscription that contains an Azure Synapse Analytics workspace named workspace1. Workspace1 contains a dedicated SQL pool named SQL Pool and an Apache Spark pool named sparkpool. Sparkpool1 contains a DataFrame named pyspark.df.

You need to write the contents of pyspark_df to a tabte in SQLPooM by using a PySpark notebook. How should you complete the code? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

Question 227
Correct answer: Question 227

You are implementing a star schema in an Azure Synapse Analytics dedicated SQL pool. You plan to create a table named DimProduct.

DimProduct must be a Type 3 slowly changing dimension (SCO) table that meets the following requirements:

• The values in two columns named ProductKey and ProductSourceID will remain the same.

• The values in three columns named ProductName, ProductDescription, and Color can change. You need to add additional columns to complete the following table definition.

A.
A.
Answers
B.
B.
Answers
C.
C.
Answers
D.
D.
Answers
E.
E.
Answers
F.
F.
Answers
Suggested answer: A, B, C

You have an Azure Data Lake Storage Gen2 account that contains two folders named Folder and Folder2.

You use Azure Data Factory to copy multiple files from Folder1 to Folder2. You receive the following error.

What should you do to resolve the error.

A.
Add an explicit mapping.
A.
Add an explicit mapping.
Answers
B.
Enable fault tolerance to skip incompatible rows.
B.
Enable fault tolerance to skip incompatible rows.
Answers
C.
Lower the degree of copy parallelism
C.
Lower the degree of copy parallelism
Answers
D.
Change the Copy activity setting to Binary Copy
D.
Change the Copy activity setting to Binary Copy
Answers
Suggested answer: A

Explanation:

Reference: https://knowledge.informatica.com/s/article/Microsoft-Azure-Data-Lake-Store-Gen2- target-file-names-not-generating-as-expected-using-create-at-runtime-option

You are designing a folder structure for the files m an Azure Data Lake Storage Gen2 account. The account has one container that contains three years of data. You need to recommend a folder structure that meets the following requirements:

• Supports partition elimination for queries by Azure Synapse Analytics serverless SQL pooh

• Supports fast data retrieval for data from the current month

• Simplifies data security management by department

Which folder structure should you recommend?

A.
\YYY\MM\DD\Department\DataSource\DataFile_YYYMMMDD.parquet
A.
\YYY\MM\DD\Department\DataSource\DataFile_YYYMMMDD.parquet
Answers
B.
\Depdftment\DataSource\YYY\MM\DataFile_YYYYMMDD.parquet
B.
\Depdftment\DataSource\YYY\MM\DataFile_YYYYMMDD.parquet
Answers
C.
\DD\MM\YYYY\Department\DataSource\DataFile_DDMMYY.parquet
C.
\DD\MM\YYYY\Department\DataSource\DataFile_DDMMYY.parquet
Answers
D.
\DataSource\Department\YYYYMM\DataFile_YYYYMMDD.parquet
D.
\DataSource\Department\YYYYMM\DataFile_YYYYMMDD.parquet
Answers
Suggested answer: B

Explanation:

Department top level in the hierarchy to simplify security management.Month (MM) at the leaf/bottom level to support fast data retrieval for data from the current month.

Total 320 questions
Go to page: of 32