exam questions

Exam Professional Machine Learning Engineer All Questions

View all questions & answers for the Professional Machine Learning Engineer exam

Exam Professional Machine Learning Engineer topic 1 question 275 discussion

Actual exam question from Google's Professional Machine Learning Engineer
Question #: 275
Topic #: 1
[All Professional Machine Learning Engineer Questions]

You have trained an XGBoost model that you plan to deploy on Vertex AI for online prediction. You are now uploading your model to Vertex AI Model Registry, and you need to configure the explanation method that will serve online prediction requests to be returned with minimal latency. You also want to be alerted when feature attributions of the model meaningfully change over time. What should you do?

  • A. 1. Specify sampled Shapley as the explanation method with a path count of 5.
    2. Deploy the model to Vertex AI Endpoints.
    3. Create a Model Monitoring job that uses prediction drift as the monitoring objective.
  • B. 1. Specify Integrated Gradients as the explanation method with a path count of 5.
    2. Deploy the model to Vertex AI Endpoints.
    3. Create a Model Monitoring job that uses prediction drift as the monitoring objective.
  • C. 1. Specify sampled Shapley as the explanation method with a path count of 50.
    2. Deploy the model to Vertex AI Endpoints.
    3. Create a Model Monitoring job that uses training-serving skew as the monitoring objective.
  • D. 1. Specify Integrated Gradients as the explanation method with a path count of 50.
    2. Deploy the model to Vertex AI Endpoints.
    3. Create a Model Monitoring job that uses training-serving skew as the monitoring objective.
Show Suggested Answer Hide Answer
Suggested Answer: A 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
daidai75
9 months, 1 week ago
Selected Answer: A
Sampled Shapley is a method suitable for XGBoost models. A lower path count (like 5) would indeed ensure lower latency in explanations, but might compromise on the precision of the explanations.Model Monitoring - Prediction Drift: This monitors the change in model predictions over time, which can indirectly indicate a change in feature attributions, but it's not directly monitoring the attributions themselves.
upvoted 3 times
...
shadz10
9 months, 2 weeks ago
Selected Answer: A
not B as integrated gradients is only for Custom-trained TensorFlow models that use a TensorFlow prebuilt container to serve predictions and AutoML image models
upvoted 3 times
shadz10
9 months, 2 weeks ago
https://cloud.google.com/vertex-ai/docs/explainable-ai/overview
upvoted 1 times
...
...
36bdc1e
9 months, 2 weeks ago
A Sampled Shapley is a fast and scalable approximation of the Shapley value, which is a game-theoretic concept that measures the contribution of each feature to the model prediction. Sampled Shapley is suitable for online prediction requests, as it can return feature attributions with minimal latency. The path count parameter controls the number of samples used to estimate the Shapley value, and a lower value means faster computation. Integrated Gradients is another explanation method that computes the average gradient along the path from a baseline input to the actual input. Integrated Gradients is more accurate than Sampled Shapley, but also more computationally intensive
upvoted 4 times
...
pikachu007
9 months, 2 weeks ago
Selected Answer: A
Explanation Method: Sampled Shapley: This method provides high-fidelity feature attributions while being computationally efficient, making it ideal for low-latency online predictions. Integrated Gradients: While also accurate, it's generally more computationally intensive than sampled Shapley, potentially introducing latency. Path Count: Lower Path Count (5): Reducing path count further decreases computation time, optimizing for faster prediction responses. Monitoring Objective: Prediction Drift: This type of monitoring detects changes in feature importance over time, aligning with the goal of tracking feature attribution shifts. Training-Serving Skew: This monitors discrepancies between training and serving data distributions, which isn't directly related to feature attributions.
upvoted 4 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago