ExamGecko
Question list
Search
Search

List of questions

Search

Related questions











Question 274 - Professional Machine Learning Engineer discussion

Report
Export

You have developed a fraud detection model for a large financial institution using Vertex AI. The model achieves high accuracy, but stakeholders are concerned about potential bias based on customer demographics. You have been asked to provide insights into the model's decision-making process and identify any fairness issues. What should you do?

A.

Enable Vertex AI Model Monitoring to detect training-serving skew. Configure an alert to send an email when the skew or drift for a model's feature exceeds a predefined threshold. Retrain the model by appending new data to existing training data.

Answers
A.

Enable Vertex AI Model Monitoring to detect training-serving skew. Configure an alert to send an email when the skew or drift for a model's feature exceeds a predefined threshold. Retrain the model by appending new data to existing training data.

B.

Compile a dataset of unfair predictions. Use Vertex AI Vector Search to identify similar data points in the model's predictions. Report these data points to the stakeholders.

Answers
B.

Compile a dataset of unfair predictions. Use Vertex AI Vector Search to identify similar data points in the model's predictions. Report these data points to the stakeholders.

C.

Use feature attribution in Vertex AI to analyze model predictions and the impact of each feature on the model's predictions.

Answers
C.

Use feature attribution in Vertex AI to analyze model predictions and the impact of each feature on the model's predictions.

D.

Create feature groups using Vertex AI Feature Store to segregate customer demographic features and non-demographic features. Retrain the model using only non-demographic features.

Answers
D.

Create feature groups using Vertex AI Feature Store to segregate customer demographic features and non-demographic features. Retrain the model using only non-demographic features.

Suggested answer: C

Explanation:

Feature attribution helps to determine how each feature influences predictions, essential for identifying bias. Vertex AI's built-in explainability tools provide insights without altering the model's feature space. Model monitoring (Option A) detects distributional drift rather than feature influence. Options B and D do not directly address the request to explain model decisions or provide fairness insights.

asked 07/11/2024
Patrick Thiel
36 questions
User
Your answer:
0 comments
Sorted by

Leave a comment first