exam questions

Exam Professional Machine Learning Engineer All Questions

View all questions & answers for the Professional Machine Learning Engineer exam

Exam Professional Machine Learning Engineer topic 1 question 245 discussion

Actual exam question from Google's Professional Machine Learning Engineer
Question #: 245
Topic #: 1
[All Professional Machine Learning Engineer Questions]

You are developing an ML model that predicts the cost of used automobiles based on data such as location, condition, model type, color, and engine/battery efficiency. The data is updated every night. Car dealerships will use the model to determine appropriate car prices. You created a Vertex AI pipeline that reads the data splits the data into training/evaluation/test sets performs feature engineering trains the model by using the training dataset and validates the model by using the evaluation dataset. You need to configure a retraining workflow that minimizes cost. What should you do?

  • A. Compare the training and evaluation losses of the current run. If the losses are similar, deploy the model to a Vertex AI endpoint. Configure a cron job to redeploy the pipeline every night.
  • B. Compare the training and evaluation losses of the current run. If the losses are similar, deploy the model to a Vertex AI endpoint with training/serving skew threshold model monitoring. When the model monitoring threshold is triggered redeploy the pipeline.
  • C. Compare the results to the evaluation results from a previous run. If the performance improved deploy the model to a Vertex AI endpoint. Configure a cron job to redeploy the pipeline every night.
  • D. Compare the results to the evaluation results from a previous run. If the performance improved deploy the model to a Vertex AI endpoint with training/serving skew threshold model monitoring. When the model monitoring threshold is triggered redeploy the pipeline.
Show Suggested Answer Hide Answer
Suggested Answer: D 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
fitri001
Highly Voted 6 months, 1 week ago
Selected Answer: D
Since the goal is to minimize cost while maintaining accuracy, Option D provides a more targeted approach for retraining based on the likelihood of the model being outdated due to data changes. Option B might trigger retraining more frequently even if the performance difference doesn't necessarily stem from a significant shift in the data distribution.
upvoted 6 times
fitri001
6 months, 1 week ago
Option D: Utilizes training/serving skew monitoring. This specifically focuses on identifying discrepancies between the training data and the real-world data the deployed model encounters. This is a strong indicator of when the model might be outdated due to changes in the data distribution. Option B: Utilizes training/serving loss monitoring. Training loss tells you how well the model performs on the training data, while serving loss tells you how well it performs on real-world data. While high serving loss can indicate a problem, it might not necessarily be due to training/serving skew. Other factors like data quality issues or concept drift (gradual changes in the underlying data patterns) could also lead to high serving loss.
upvoted 3 times
...
...
guilhermebutzke
Highly Voted 8 months, 2 weeks ago
Selected Answer: D
My answer D: A and C: Not Correct: Schedule a retrain every night is not necessary since the model is performing well. B. Not Correct: This approach focuses on internal consistency within the current training run, train versus loss evaluation. Comparing similar training and validation losses doesn't guarantee better performance than previous models. This is an approach to identity overfitting, for example, or model quality. D. Correct: This approach focuses on identifying performance changes over time. Comparing to previous runs helps assess if the new model performs better than the old one on the evaluation set. we will check if this new version is better or not than the old one https://www.youtube.com/watch?v=1ykDWsnL2LE&ab_channel=GoogleCloudTech
upvoted 5 times
...
omermahgoub
Most Recent 6 months, 2 weeks ago
Selected Answer: D
D. Compare the results to the evaluation results from a previous run. If the performance improved, deploy the model to a Vertex AI endpoint with training/serving skew threshold model monitoring. When the model monitoring threshold is triggered, redeploy the pipeline.
upvoted 3 times
pinimichele01
6 months, 1 week ago
i agree, see guilhermebutzke
upvoted 1 times
...
...
pikachu007
9 months, 2 weeks ago
Selected Answer: B
Option A: Redeploying the pipeline every night without checking for degradation wastes resources if model performance is stable. Option C: Comparing results to a previous run doesn't guarantee model degradation detection in the current run. Option D: Comparing to a previous run and using model monitoring is redundant; model monitoring alone is sufficient.
upvoted 5 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago