ExamGecko

Microsoft DP-100 Practice Test - Questions Answers, Page 6

Question list
Search
Search

List of questions

Search

Related questions











DRAG DROP

You have a model with a large difference between the training and validation error values.

You must create a new model and perform cross-validation.

You need to identify a parameter set for the new model using Azure Machine Learning Studio.

Which module you should use for each step? To answer, drag the appropriate modules to the correct steps. Each module may be used once or more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.

NOTE: Each correct selection is worth one point.


Question 51
Correct answer: Question 51

Explanation:

Box 1: Split data

Box 2: Partition and Sample

Box 3: Two-Class Boosted Decision Tree

Box 4: Tune Model Hyperparameters Integrated train and tune: You configure a set of parameters to use, and then let the module iterate over multiple combinations, measuring accuracy until it finds a "best" model. With most learner modules, you can choose which parameters should be changed during the training process, and which should remain fixed.

We recommend that you use Cross-Validate Model to establish the goodness of the model given the specified parameters. Use Tune Model Hyperparameters to identify the optimal parameters.

References:

https://docs.microsoft.com/en-us/azure/machine-learning/studio-module-reference/partition-and-sample

HOTSPOT

You are analyzing the asymmetry in a statistical distribution.

The following image contains two density curves that show the probability distribution of two datasets.

Use the drop-down menus to select the answer choice that answers each question based on the information presented in the graphic.

NOTE: Each correct selection is worth one point.


Question 52
Correct answer: Question 52

Explanation:

Box 1: Positive skew

Positive skew values means the distribution is skewed to the right.

Box 2: Negative skew

Negative skewness values mean the distribution is skewed to the left.

References:

https://docs.microsoft.com/en-us/azure/machine-learning/studio-module-reference/compute-elementary-statistics

HOTSPOT

You train a classification model by using a decision tree algorithm.

You create an estimator by running the following Python code. The variable feature_names is a list of all feature names, and class_names is a list of all class names.

from interpret.ext.blackbox import TabularExplainer

You need to explain the predictions made by the model for all classes by determining the importance of all features.

For each of the following statements, select Yes if the statement is true. Otherwise, select No.

NOTE: Each correct selection is worth one point.


Question 53
Correct answer: Question 53

Explanation:

Box 1: Yes

TabularExplainer calls one of the three SHAP explainers underneath (TreeExplainer, DeepExplainer, or KernelExplainer).

Box 2: Yes

To make your explanations and visualizations more informative, you can choose to pass in feature names and output class names if doing classification.

Box 3: No TabularExplainer automatically selects the most appropriate one for your use case, but you can call each of its three underlying explainers underneath (TreeExplainer, DeepExplainer, or KernelExplainer) directly.

Reference:

https://docs.microsoft.com/en-us/azure/machine-learning/how-to-machine-learning-interpretability-aml

DRAG DROP

You have several machine learning models registered in an Azure Machine Learning workspace.

You must use the Fairlearn dashboard to assess fairness in a selected model.

Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.


Question 54
Correct answer: Question 54

Explanation:

Step 1: Select a model feature to be evaluated.

Step 2: Select a binary classification or regression model.

Register your models within Azure Machine Learning. For convenience, store the results in a dictionary, which maps the id of the registered model (a string in name:version format) to the predictor itself.

Example:

model_dict = {}

lr_reg_id = register_model("fairness_logistic_regression", lr_predictor)

model_dict[lr_reg_id] = lr_predictor

svm_reg_id = register_model("fairness_svm", svm_predictor)

model_dict[svm_reg_id] = svm_predictor

Step 3: Select a metric to be measured

Precompute fairness metrics.

Create a dashboard dictionary using Fairlearn's metrics package.

Reference:

https://docs.microsoft.com/en-us/azure/machine-learning/how-to-machine-learning-fairness-aml

HOTSPOT

A biomedical research company plans to enroll people in an experimental medical treatment trial.

You create and train a binary classification model to support selection and admission of patients to the trial. The model includes the following features: Age, Gender, and Ethnicity.

The model returns different performance metrics for people from different ethnic groups.

You need to use Fairlearn to mitigate and minimize disparities for each category in the Ethnicity feature.

Which technique and constraint should you use? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.


Question 55
Correct answer: Question 55

Explanation:

Box 1: Grid Search

Fairlearn open-source package provides postprocessing and reduction unfairness mitigation algorithms: ExponentiatedGradient, GridSearch, and ThresholdOptimizer.

Note: The Fairlearn open-source package provides postprocessing and reduction unfairness mitigation algorithms types:

Reduction: These algorithms take a standard black-box machine learning estimator (e.g., a LightGBM model) and generate a set of retrained models using a sequence of re-weighted training datasets.

Post-processing: These algorithms take an existing classifier and the sensitive feature as input.

Box 2: Demographic parity

The Fairlearn open-source package supports the following types of parity constraints: Demographic parity, Equalized odds, Equal opportunity, and Bounded group loss.

Reference:

https://docs.microsoft.com/en-us/azure/machine-learning/concept-fairness-ml

You are developing a hands-on workshop to introduce Docker for Windows to attendees.

You need to ensure that workshop attendees can install Docker on their devices.

Which two prerequisite components should attendees install on the devices? Each correct answer presents part of the solution.

NOTE: Each correct selection is worth one point.

A.
Microsoft Hardware-Assisted Virtualization Detection Tool
A.
Microsoft Hardware-Assisted Virtualization Detection Tool
Answers
B.
Kitematic
B.
Kitematic
Answers
C.
BIOS-enabled virtualization
C.
BIOS-enabled virtualization
Answers
D.
VirtualBox
D.
VirtualBox
Answers
E.
Windows 10 64-bit Professional
E.
Windows 10 64-bit Professional
Answers
Suggested answer: C, E

Explanation:

C: Make sure your Windows system supports Hardware Virtualization Technology and that virtualization is enabled.

Ensure that hardware virtualization support is turned on in the BIOS settings. For example:

E: To run Docker, your machine must have a 64-bit operating system running Windows 7 or higher.

Reference:

https://docs.docker.com/toolbox/toolbox_install_windows/

https://blogs.technet.microsoft.com/canitpro/2015/09/08/step-by-step-enabling-hyper-v-for-use-on-windows-10/

Your team is building a data engineering and data science development environment.

The environment must support the following requirements:

support Python and Scala

compose data storage, movement, and processing services into automated data pipelines

the same tool should be used for the orchestration of both data engineering and data science

support workload isolation and interactive workloads

enable scaling across a cluster of machines

You need to create the environment.

What should you do?

A.
Build the environment in Apache Hive for HDInsight and use Azure Data Factory for orchestration.
A.
Build the environment in Apache Hive for HDInsight and use Azure Data Factory for orchestration.
Answers
B.
Build the environment in Azure Databricks and use Azure Data Factory for orchestration.
B.
Build the environment in Azure Databricks and use Azure Data Factory for orchestration.
Answers
C.
Build the environment in Apache Spark for HDInsight and use Azure Container Instances for orchestration.
C.
Build the environment in Apache Spark for HDInsight and use Azure Container Instances for orchestration.
Answers
D.
Build the environment in Azure Databricks and use Azure Container Instances for orchestration.
D.
Build the environment in Azure Databricks and use Azure Container Instances for orchestration.
Answers
Suggested answer: B

Explanation:

In Azure Databricks, we can create two different types of clusters.

Standard, these are the default clusters and can be used with Python, R, Scala and SQL High-concurrency

Azure Databricks is fully integrated with Azure Data Factory.

Incorrect Answers:

D: Azure Container Instances is good for development or testing. Not suitable for production workloads.

Reference: https://docs.microsoft.com/en-us/azure/architecture/data-guide/technology-choices/data-science-and-machine-learning

You plan to build a team data science environment. Data for training models in machine learning pipelines will be over 20 GB in size.

You have the following requirements:

Models must be built using Caffe2 or Chainer frameworks.

Data scientists must be able to use a data science environment to build the machine learning pipelines and train models on their personal devices in both connected and disconnected network environments.

Personal devices must support updating machine learning pipelines when connected to a network.

You need to select a data science environment.

Which environment should you use?

A.
Azure Machine Learning Service
A.
Azure Machine Learning Service
Answers
B.
Azure Machine Learning Studio
B.
Azure Machine Learning Studio
Answers
C.
Azure Databricks
C.
Azure Databricks
Answers
D.
Azure Kubernetes Service (AKS)
D.
Azure Kubernetes Service (AKS)
Answers
Suggested answer: A

Explanation:

The Data Science Virtual Machine (DSVM) is a customized VM image on Microsoft's Azure cloud built specifically for doing data science. Caffe2 and Chainer are supported by DSVM. DSVM integrates with Azure Machine Learning.

Incorrect Answers:

B: Use Machine Learning Studio when you want to experiment with machine learning models quickly and easily, and the built-in machine learning algorithms are sufficient for your solutions.

Reference: https://docs.microsoft.com/en-us/azure/machine-learning/data-science-virtual-machine/overview

You are implementing a machine learning model to predict stock prices.

The model uses a PostgreSQL database and requires GPU processing.

You need to create a virtual machine that is pre-configured with the required tools.

What should you do?

A.
Create a Data Science Virtual Machine (DSVM) Windows edition.
A.
Create a Data Science Virtual Machine (DSVM) Windows edition.
Answers
B.
Create a Geo Al Data Science Virtual Machine (Geo-DSVM) Windows edition.
B.
Create a Geo Al Data Science Virtual Machine (Geo-DSVM) Windows edition.
Answers
C.
Create a Deep Learning Virtual Machine (DLVM) Linux edition.
C.
Create a Deep Learning Virtual Machine (DLVM) Linux edition.
Answers
D.
Create a Deep Learning Virtual Machine (DLVM) Windows edition.
D.
Create a Deep Learning Virtual Machine (DLVM) Windows edition.
Answers
Suggested answer: A

Explanation:

In the DSVM, your training models can use deep learning algorithms on hardware that's based on graphics processing units (GPUs).

PostgreSQL is available for the following operating systems: Linux (all recent distributions), 64-bit installers available for macOS (OS X) version 10.6 and newer - Windows (with installers available for 64-bit version; tested on latest versions and back to Windows 2012 R2.

Incorrect Answers:

B: The Azure Geo AI Data Science VM (Geo-DSVM) delivers geospatial analytics capabilities from Microsoft's Data Science VM. Specifically, this VM extends the AI and data science toolkits in the Data Science VM by adding ESRI's market-leading ArcGIS Pro Geographic Information System.

C, D: DLVM is a template on top of DSVM image. In terms of the packages, GPU drivers etc are all there in the DSVM image. Mostly it is for convenience during creation where we only allow DLVM to be created on GPU VM instances on

Azure.

Reference: https://docs.microsoft.com/en-us/azure/machine-learning/data-science-virtual-machine/overview

You are developing deep learning models to analyze semi-structured, unstructured, and structured data types.

You have the following data available for model building:

Video recordings of sporting events

Transcripts of radio commentary about events

Logs from related social media feeds captured during sporting events

You need to select an environment for creating the model.

Which environment should you use?

A.
Azure Cognitive Services
A.
Azure Cognitive Services
Answers
B.
Azure Data Lake Analytics
B.
Azure Data Lake Analytics
Answers
C.
Azure HDInsight with Spark MLib
C.
Azure HDInsight with Spark MLib
Answers
D.
Azure Machine Learning Studio
D.
Azure Machine Learning Studio
Answers
Suggested answer: A

Explanation:

Azure Cognitive Services expand on Microsoft's evolving portfolio of machine learning APIs and enable developers to easily add cognitive features - such as emotion and video detection; facial, speech, and vision recognition; and speech and language understanding - into their applications. The goal of Azure Cognitive Services is to help developers create applications that can see, hear, speak, understand, and even begin to reason. The catalog of services within Azure

Cognitive Services can be categorized into five main pillars - Vision, Speech, Language, Search, and Knowledge.

Reference: https://docs.microsoft.com/en-us/azure/cognitive-services/welcome

Total 433 questions
Go to page: of 44