ExamGecko
Question list
Search
Search

List of questions

Search

Related questions











Question 13 - Professional Machine Learning Engineer discussion

Report
Export

You work for a global footwear retailer and need to predict when an item will be out of stock based on historical inventory data. Customer behavior is highly dynamic since footwear demand is influenced by many different factors. You want to serve models that are trained on all available data, but track your performance on specific subsets of data before pushing to production. What is the most streamlined and reliable way to perform this validation?

A.
Use the TFX ModelValidator tools to specify performance metrics for production readiness
Answers
A.
Use the TFX ModelValidator tools to specify performance metrics for production readiness
B.
Use k-fold cross-validation as a validation strategy to ensure that your model is ready for production.
Answers
B.
Use k-fold cross-validation as a validation strategy to ensure that your model is ready for production.
C.
Use the last relevant week of data as a validation set to ensure that your model is performing accurately on current data
Answers
C.
Use the last relevant week of data as a validation set to ensure that your model is performing accurately on current data
D.
Use the entire dataset and treat the area under the receiver operating characteristics curve (AUC ROC) as the main metric.
Answers
D.
Use the entire dataset and treat the area under the receiver operating characteristics curve (AUC ROC) as the main metric.
Suggested answer: A

Explanation:

TFX ModelValidatoris a tool that allows you to compare new models against a baseline model and evaluate their performance on different metrics and data slices1. You can use this tool to validate your models before deploying them to production and ensure that they meet your expectations and requirements.

k-fold cross-validationis a technique that splits the data into k subsets and trains the model on k-1 subsets while testing it on the remaining subset.This is repeated k times and the average performance is reported2. This technique is useful for estimating the generalization error of a model, but it does not account for the dynamic nature of customer behavior or the potential changes in data distribution over time.

Using the last relevant week of data as a validation setis a simple way to check the model's performance on recent data, but it may not be representative of the entire data or capture the long-term trends and patterns. It also does not allow you to compare the model with a baseline or evaluate it on different data slices.

Using the entire dataset and treating the AUC ROC as the main metricis not a good practice because it does not leave any data for validation or testing. It also assumes that the AUC ROC is the only metric that matters, which may not be true for your business problem. You may want to consider other metrics such as precision, recall, or revenue.

asked 18/09/2024
JENNIFER MALIWANAG
38 questions
User
Your answer:
0 comments
Sorted by

Leave a comment first