ExamGecko
Home Home / Google / Professional Machine Learning Engineer

Google Professional Machine Learning Engineer Practice Test - Questions Answers, Page 22

Question list
Search
Search

List of questions

Search

Related questions











Your team is training a large number of ML models that use different algorithms, parameters and datasets. Some models are trained in Vertex Ai Pipelines, and some are trained on Vertex Al Workbench notebook instances. Your team wants to compare the performance of the models across both services. You want to minimize the effort required to store the parameters and metrics What should you do?

A.
Implement an additional step for all the models running in pipelines and notebooks to export parameters and metrics to BigQuery.
A.
Implement an additional step for all the models running in pipelines and notebooks to export parameters and metrics to BigQuery.
Answers
B.
Create a Vertex Al experiment Submit all the pipelines as experiment runs. For models trained on notebooks log parameters and metrics by using the Vertex Al SDK.
B.
Create a Vertex Al experiment Submit all the pipelines as experiment runs. For models trained on notebooks log parameters and metrics by using the Vertex Al SDK.
Answers
C.
Implement all models in Vertex Al Pipelines Create a Vertex Al experiment, and associate all pipeline runs with that experiment.
C.
Implement all models in Vertex Al Pipelines Create a Vertex Al experiment, and associate all pipeline runs with that experiment.
Answers
D.
Store all model parameters and metrics as mode! metadata by using the Vertex Al Metadata API.
D.
Store all model parameters and metrics as mode! metadata by using the Vertex Al Metadata API.
Answers
Suggested answer: B

Explanation:

Vertex AI Experiments is a service that allows you to track, compare, and manage experiments with Vertex AI. You can use Vertex AI Experiments to record the parameters, metrics, and artifacts of each model training run, and compare them in a graphical interface. Vertex AI Experiments supports models trained in Vertex AI Pipelines, Vertex AI Custom Training, and Vertex AI Workbench notebooks. To use Vertex AI Experiments, you need to create an experiment and submit your pipeline runs or custom training jobs as experiment runs. For models trained on notebooks, you need to use the Vertex AI SDK to log the parameters and metrics to the experiment. This way, you can minimize the effort required to store and compare the model performance across different services.Reference:Track, compare, manage experiments with Vertex AI Experiments,Vertex AI Pipelines: Metrics visualization and run comparison using the KFP SDK, [Vertex AI SDK for Python]

You work on a team that builds state-of-the-art deep learning models by using the TensorFlow framework. Your team runs multiple ML experiments each week which makes it difficult to track the experiment runs. You want a simple approach to effectively track, visualize and debug ML experiment runs on Google Cloud while minimizing any overhead code. How should you proceed?

A.
Set up Vertex Al Experiments to track metrics and parameters Configure Vertex Al TensorBoard for visualization.
A.
Set up Vertex Al Experiments to track metrics and parameters Configure Vertex Al TensorBoard for visualization.
Answers
B.
Set up a Cloud Function to write and save metrics files to a Cloud Storage Bucket Configure a Google Cloud VM to host TensorBoard locally for visualization.
B.
Set up a Cloud Function to write and save metrics files to a Cloud Storage Bucket Configure a Google Cloud VM to host TensorBoard locally for visualization.
Answers
C.
Set up a Vertex Al Workbench notebook instance Use the instance to save metrics data in a Cloud Storage bucket and to host TensorBoard locally for visualization.
C.
Set up a Vertex Al Workbench notebook instance Use the instance to save metrics data in a Cloud Storage bucket and to host TensorBoard locally for visualization.
Answers
D.
Set up a Cloud Function to write and save metrics files to a BigQuery table. Configure a Google Cloud VM to host TensorBoard locally for visualization.
D.
Set up a Cloud Function to write and save metrics files to a BigQuery table. Configure a Google Cloud VM to host TensorBoard locally for visualization.
Answers
Suggested answer: A

Explanation:

Vertex AI Experiments is a service that allows you to track, compare, and optimize your ML experiments on Google Cloud. You can use Vertex AI Experiments to log metrics and parameters from your TensorFlow models, and then visualize them in Vertex AI TensorBoard. Vertex AI TensorBoard is a managed service that provides a web interface for viewing and debugging your ML experiments. You can use Vertex AI TensorBoard to compare different runs, inspect model graphs, analyze scalars, histograms, images, and more. By using Vertex AI Experiments and Vertex AI TensorBoard, you can simplify your ML experiment tracking and visualization workflow, and avoid the overhead of setting up and maintaining your own Cloud Functions, Cloud Storage buckets, or VMs.Reference:

[Vertex AI Experiments documentation]

[Vertex AI TensorBoard documentation]

Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

Your work for a textile manufacturing company. Your company has hundreds of machines and each machine has many sensors. Your team used the sensory data to build hundreds of ML models that detect machine anomalies Models are retrained daily and you need to deploy these models in a cost-effective way. The models must operate 24/7 without downtime and make sub millisecond predictions. What should you do?

A.
Deploy a Dataflow batch pipeline and a Vertex Al Prediction endpoint.
A.
Deploy a Dataflow batch pipeline and a Vertex Al Prediction endpoint.
Answers
B.
Deploy a Dataflow batch pipeline with the Runlnference API. and use model refresh.
B.
Deploy a Dataflow batch pipeline with the Runlnference API. and use model refresh.
Answers
C.
Deploy a Dataflow streaming pipeline and a Vertex Al Prediction endpoint with autoscaling.
C.
Deploy a Dataflow streaming pipeline and a Vertex Al Prediction endpoint with autoscaling.
Answers
D.
Deploy a Dataflow streaming pipeline with the Runlnference API and use automatic model refresh.
D.
Deploy a Dataflow streaming pipeline with the Runlnference API and use automatic model refresh.
Answers
Suggested answer: D

Explanation:

A Dataflow streaming pipeline is a cost-effective way to process large volumes of real-time data from sensors. The RunInference API is a Dataflow transform that allows you to run online predictions on your streaming data using your ML models. By using the RunInference API, you can avoid the latency and cost of using a separate prediction service. The automatic model refresh feature enables you to update your models in the pipeline without redeploying the pipeline. This way, you can ensure that your models are always up-to-date and accurate. By deploying a Dataflow streaming pipeline with the RunInference API and using automatic model refresh, you can achieve sub-millisecond predictions, 24/7 availability, and low operational overhead for your ML models.Reference:

Dataflow documentation

RunInference API documentation

Automatic model refresh documentation

Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

You are developing an ML model that predicts the cost of used automobiles based on data such as location, condition model type color, and engine-'battery efficiency. The data is updated every night Car dealerships will use the model to determine appropriate car prices. You created a Vertex Al pipeline that reads the data splits the data into training/evaluation/test sets performs feature engineering trains the model by using the training dataset and validates the model by using the evaluation dataset. You need to configure a retraining workflow that minimizes cost What should you do?

A.
Compare the training and evaluation losses of the current run If the losses are similar, deploy the model to a Vertex AI endpoint Configure a cron job to redeploy the pipeline every night.
A.
Compare the training and evaluation losses of the current run If the losses are similar, deploy the model to a Vertex AI endpoint Configure a cron job to redeploy the pipeline every night.
Answers
B.
Compare the training and evaluation losses of the current run If the losses are similar deploy the model to a Vertex Al endpoint with training/serving skew threshold model monitoring When the model monitoring threshold is tnggered redeploy the pipeline.
B.
Compare the training and evaluation losses of the current run If the losses are similar deploy the model to a Vertex Al endpoint with training/serving skew threshold model monitoring When the model monitoring threshold is tnggered redeploy the pipeline.
Answers
C.
Compare the results to the evaluation results from a previous run If the performance improved deploy the model to a Vertex Al endpoint Configure a cron job to redeploy the pipeline every night.
C.
Compare the results to the evaluation results from a previous run If the performance improved deploy the model to a Vertex Al endpoint Configure a cron job to redeploy the pipeline every night.
Answers
D.
Compare the results to the evaluation results from a previous run If the performance improved deploy the model to a Vertex Al endpoint with training/serving skew threshold model monitoring. When the model monitoring threshold is triggered, redeploy the pipeline.
D.
Compare the results to the evaluation results from a previous run If the performance improved deploy the model to a Vertex Al endpoint with training/serving skew threshold model monitoring. When the model monitoring threshold is triggered, redeploy the pipeline.
Answers
Suggested answer: B

Explanation:

Comparing the training and evaluation losses of the current run is a good way to check if the model is overfitting or underfitting. If the losses are similar, it means that the model is generalizing well and can be deployed to a Vertex AI endpoint. Vertex AI endpoint is a service that allows you to serve your ML models online and scale them automatically. By using a training/serving skew threshold model monitoring, you can detect if there is a significant difference between the data used for training and the data used for serving. This can indicate that the model is becoming stale or inaccurate over time. When the model monitoring threshold is triggered, it means that the model needs to be retrained with the latest data. By redeploying the pipeline, you can automate the retraining process and update the model with the new data. This way, you can minimize the cost of retraining and ensure that your model is always up-to-date and accurate.Reference:

Vertex AI documentation

Vertex AI endpoint documentation

Model monitoring documentation

Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

You recently used BigQuery ML to train an AutoML regression model. You shared results with your team and received positive feedback. You need to deploy your model for online prediction as quickly as possible. What should you do?

A.
Retrain the model by using BigQuery ML. and specify Vertex Al as the model registry Deploy the model from Vertex Al Model Registry to a Vertex Al endpoint.
A.
Retrain the model by using BigQuery ML. and specify Vertex Al as the model registry Deploy the model from Vertex Al Model Registry to a Vertex Al endpoint.
Answers
B.
Retrain the model by using Vertex Al Deploy the model from Vertex Al Model Registry to a Vertex Al endpoint.
B.
Retrain the model by using Vertex Al Deploy the model from Vertex Al Model Registry to a Vertex Al endpoint.
Answers
C.
Alter the model by using BigQuery ML and specify Vertex Al as the model registry Deploy the model from Vertex Al Model Registry to a Vertex Al endpoint.
C.
Alter the model by using BigQuery ML and specify Vertex Al as the model registry Deploy the model from Vertex Al Model Registry to a Vertex Al endpoint.
Answers
D.
Export the model from BigQuery ML to Cloud Storage Import the model into Vertex Al Model Registry Deploy the model to a Vertex Al endpoint.
D.
Export the model from BigQuery ML to Cloud Storage Import the model into Vertex Al Model Registry Deploy the model to a Vertex Al endpoint.
Answers
Suggested answer: A

Explanation:

BigQuery ML is a service that allows you to create and train ML models using SQL queries. You can use BigQuery ML to train an AutoML regression model, which is a type of model that automatically selects the best features and architecture for your data. You can also specify Vertex AI as the model registry, which is a service that allows you to store and manage your ML models. By using Vertex AI as the model registry, you can easily deploy your model to a Vertex AI endpoint, which is a service that allows you to serve your ML models online and scale them automatically. By using BigQuery ML, Vertex AI model registry, and Vertex AI endpoint, you can deploy your model for online prediction as quickly as possible, without having to export, import, or retrain your model.Reference:

BigQuery ML documentation

Vertex AI documentation

Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

You built a deep learning-based image classification model by using on-premises data. You want to use Vertex Al to deploy the model to production Due to security concerns you cannot move your data to the cloud. You are aware that the input data distribution might change over time You need to detect model performance changes in production. What should you do?

A.
Use Vertex Explainable Al for model explainability Configure feature-based explanations.
A.
Use Vertex Explainable Al for model explainability Configure feature-based explanations.
Answers
B.
Use Vertex Explainable Al for model explainability Configure example-based explanations.
B.
Use Vertex Explainable Al for model explainability Configure example-based explanations.
Answers
C.
Create a Vertex Al Model Monitoring job. Enable training-serving skew detection for your model.
C.
Create a Vertex Al Model Monitoring job. Enable training-serving skew detection for your model.
Answers
D.
Create a Vertex Al Model Monitoring job. Enable feature attribution skew and dnft detection for your model.
D.
Create a Vertex Al Model Monitoring job. Enable feature attribution skew and dnft detection for your model.
Answers
Suggested answer: C

Explanation:

Vertex AI Model Monitoring is a service that allows you to monitor the performance and quality of your ML models in production. You can use Vertex AI Model Monitoring to detect changes in the input data distribution, the prediction output distribution, or the model accuracy over time. Training-serving skew detection is a feature of Vertex AI Model Monitoring that compares the statistics of the data used for training the model and the data used for serving the model. If there is a significant difference between the two data distributions, it indicates that the model might be outdated or inaccurate. By enabling training-serving skew detection for your model, you can detect model performance changes in production and trigger retraining or redeployment of your model as needed. This way, you can ensure that your model is always up-to-date and accurate, without moving your data to the cloud.Reference:

Vertex AI Model Monitoring documentation

Training-serving skew detection documentation

Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

You trained a model, packaged it with a custom Docker container for serving, and deployed it to Vertex Al Model Registry. When you submit a batch prediction job, it fails with this error 'Error model server never became ready Please validate that your model file or container configuration are valid. There are no additional errors in the logs What should you do?

A.
Add a logging configuration to your application to emit logs to Cloud Logging.
A.
Add a logging configuration to your application to emit logs to Cloud Logging.
Answers
B.
Change the HTTP port in your model's configuration to the default value of 8080
B.
Change the HTTP port in your model's configuration to the default value of 8080
Answers
C.
Change the health Route value in your models configuration to /heal thcheck.
C.
Change the health Route value in your models configuration to /heal thcheck.
Answers
D.
Pull the Docker image locally and use the decker run command to launch it locally. Use the docker logs command to explore the error logs.
D.
Pull the Docker image locally and use the decker run command to launch it locally. Use the docker logs command to explore the error logs.
Answers
Suggested answer: B

Explanation:

When you deploy a custom container to Vertex AI Model Registry, you need to follow some requirements for the container configuration. One of these requirements is to use the HTTP port 8080 for serving predictions. If you use a different port, the model server might not be able to communicate with Vertex AI and cause the error ''Error model server never became ready''. To fix this error, you need to change the HTTP port in your model's configuration to the default value of 8080 and redeploy the container.Reference:

Custom container requirements documentation

Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

You are developing an ML model to identify your company s products in images. You have access to over one million images in a Cloud Storage bucket. You plan to experiment with different TensorFlow models by using Vertex Al Training You need to read images at scale during training while minimizing data I/O bottlenecks What should you do?

A.
Load the images directly into the Vertex Al compute nodes by using Cloud Storage FUSE Read the images by using the tf .data.Dataset.from_tensor_slices function.
A.
Load the images directly into the Vertex Al compute nodes by using Cloud Storage FUSE Read the images by using the tf .data.Dataset.from_tensor_slices function.
Answers
B.
Create a Vertex Al managed dataset from your image data Access the aip_training_data_uri environment variable to read the images by using the tf. data. Dataset. Iist_flies function.
B.
Create a Vertex Al managed dataset from your image data Access the aip_training_data_uri environment variable to read the images by using the tf. data. Dataset. Iist_flies function.
Answers
C.
Convert the images to TFRecords and store them in a Cloud Storage bucket Read the TFRecords by using the tf. ciata.TFRecordDataset function.
C.
Convert the images to TFRecords and store them in a Cloud Storage bucket Read the TFRecords by using the tf. ciata.TFRecordDataset function.
Answers
D.
Store the URLs of the images in a CSV file Read the file by using the tf.data.experomental.CsvDataset function.
D.
Store the URLs of the images in a CSV file Read the file by using the tf.data.experomental.CsvDataset function.
Answers
Suggested answer: C

Explanation:

TFRecords are a binary file format that can store large amounts of data efficiently. By converting the images to TFRecords and storing them in a Cloud Storage bucket, you can reduce the data size and improve the data transfer speed. You can then read the TFRecords by using the tf.data.TFRecordDataset function, which creates a dataset of tensors from the TFRecord files. This way, you can read images at scale during training while minimizing data I/O bottlenecks.Reference:

TFRecord documentation

tf.data.TFRecordDataset documentation

Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

You work at an ecommerce startup. You need to create a customer churn prediction model Your company's recent sales records are stored in a BigQuery table You want to understand how your initial model is making predictions. You also want to iterate on the model as quickly as possible while minimizing cost How should you build your first model?

A.
Export the data to a Cloud Storage Bucket Load the data into a pandas DataFrame on Vertex Al Workbench and train a logistic regression model with scikit-learn.
A.
Export the data to a Cloud Storage Bucket Load the data into a pandas DataFrame on Vertex Al Workbench and train a logistic regression model with scikit-learn.
Answers
B.
Create a tf.data.Dataset by using the TensorFlow BigQueryChent Implement a deep neural network in TensorFlow.
B.
Create a tf.data.Dataset by using the TensorFlow BigQueryChent Implement a deep neural network in TensorFlow.
Answers
C.
Prepare the data in BigQuery and associate the data with a Vertex Al dataset Create an AutoMLTabuiarTrainmgJob to train a classification model.
C.
Prepare the data in BigQuery and associate the data with a Vertex Al dataset Create an AutoMLTabuiarTrainmgJob to train a classification model.
Answers
D.
Export the data to a Cloud Storage Bucket Create tf. data. Dataset to read the data from Cloud Storage Implement a deep neural network in TensorFlow.
D.
Export the data to a Cloud Storage Bucket Create tf. data. Dataset to read the data from Cloud Storage Implement a deep neural network in TensorFlow.
Answers
Suggested answer: C

Explanation:

BigQuery is a service that allows you to store and query large amounts of data in a scalable and cost-effective way. You can use BigQuery to prepare the data for your customer churn prediction model, such as filtering, aggregating, and transforming the data. You can then associate the data with a Vertex AI dataset, which is a service that allows you to store and manage your ML data on Google Cloud. By using a Vertex AI dataset, you can easily access the data from other Vertex AI services, such as AutoML. AutoML is a service that allows you to create and train ML models without writing code. You can use AutoML to create an AutoMLTabularTrainingJob, which is a type of job that trains a classification model for tabular data, such as customer churn. By using an AutoMLTabularTrainingJob, you can benefit from the automated feature engineering, model selection, and hyperparameter tuning that AutoML provides. You can also use Vertex Explainable AI to understand how your model is making predictions, such as which features are most important and how they affect the prediction outcome. By using BigQuery, Vertex AI dataset, and AutoMLTabularTrainingJob, you can build your first model as quickly as possible while minimizing cost and complexity.Reference:

BigQuery documentation

Vertex AI dataset documentation

AutoMLTabularTrainingJob documentation

Vertex Explainable AI documentation

Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

You are developing a training pipeline for a new XGBoost classification model based on tabular data The data is stored in a BigQuery table You need to complete the following steps

1. Randomly split the data into training and evaluation datasets in a 65/35 ratio

2. Conduct feature engineering

3 Obtain metrics for the evaluation dataset.

4 Compare models trained in different pipeline executions

How should you execute these steps'?

A.
1 Using Vertex Al Pipelines, add a component to divide the data into training and evaluation sets, and add another component for feature engineering 2. Enable auto logging of metrics in the training component. 3 Compare pipeline runs in Vertex Al Experiments
A.
1 Using Vertex Al Pipelines, add a component to divide the data into training and evaluation sets, and add another component for feature engineering 2. Enable auto logging of metrics in the training component. 3 Compare pipeline runs in Vertex Al Experiments
Answers
B.
1 Using Vertex Al Pipelines, add a component to divide the data into training and evaluation sets, and add another component for feature engineering 2 Enable autologging of metrics in the training component 3 Compare models using the artifacts lineage in Vertex ML Metadata
B.
1 Using Vertex Al Pipelines, add a component to divide the data into training and evaluation sets, and add another component for feature engineering 2 Enable autologging of metrics in the training component 3 Compare models using the artifacts lineage in Vertex ML Metadata
Answers
C.
1 In BigQuery ML. use the create model statement with bocstzd_tree_classifier as the model type and use BigQuery to handle the data splits. 2 Use a SQL view to apply feature engineering and train the model using the data in that view 3. Compare the evaluation metrics of the models by using a SQL query with the ml. training_infc statement.
C.
1 In BigQuery ML. use the create model statement with bocstzd_tree_classifier as the model type and use BigQuery to handle the data splits. 2 Use a SQL view to apply feature engineering and train the model using the data in that view 3. Compare the evaluation metrics of the models by using a SQL query with the ml. training_infc statement.
Answers
D.
1 In BigQuery ML use the create model statement with boosted_tree_classifier as the model type, and use BigQuery to handle the data splits. 2 Use ml transform to specify the feature engineering transformations, and train the model using the data in the table ' 3. Compare the evaluation metrics of the models by using a SQL query with the ml. training_info statement.
D.
1 In BigQuery ML use the create model statement with boosted_tree_classifier as the model type, and use BigQuery to handle the data splits. 2 Use ml transform to specify the feature engineering transformations, and train the model using the data in the table ' 3. Compare the evaluation metrics of the models by using a SQL query with the ml. training_info statement.
Answers
Suggested answer: B

Explanation:

Vertex AI Pipelines is a service that allows you to create and run scalable and portable ML pipelines on Google Cloud. You can use Vertex AI Pipelines to add a component to divide the data into training and evaluation sets, and add another component for feature engineering. A component is a self-contained piece of code that performs a specific task in the pipeline. You can use the built-in components provided by Vertex AI Pipelines, or create your own custom components. By using Vertex AI Pipelines, you can orchestrate and automate your ML workflow, and track the provenance and lineage of your data and models. You can also enable autologging of metrics in the training component, which is a feature that automatically logs the metrics from your XGBoost model to Vertex AI Experiments. Vertex AI Experiments is a service that allows you to track, compare, and optimize your ML experiments on Google Cloud. You can use Vertex AI Experiments to monitor the training progress, visualize the metrics, and analyze the results of your model. You can also compare models using the artifacts lineage in Vertex ML Metadata. Vertex ML Metadata is a service that stores and manages the metadata of your ML artifacts, such as datasets, models, metrics, and executions. You can use Vertex ML Metadata to view the artifacts lineage, which is a graph that shows the relationships and dependencies among the artifacts. By using the artifacts lineage, you can compare the performance and quality of different models trained in different pipeline executions, and identify the best model for your use case. By using Vertex AI Pipelines, Vertex AI Experiments, and Vertex ML Metadata, you can execute the steps required for developing a training pipeline for a new XGBoost classification model based on tabular data stored in a BigQuery table.Reference:

Vertex AI Pipelines documentation

Vertex AI Experiments documentation

Vertex ML Metadata documentation

Preparing for Google Cloud Certification: Machine Learning Engineer Professional Certificate

Total 285 questions
Go to page: of 29