Welcome to ExamTopics
ExamTopics Logo
- Expert Verified, Online, Free.
exam questions

Exam Professional Machine Learning Engineer All Questions

View all questions & answers for the Professional Machine Learning Engineer exam

Exam Professional Machine Learning Engineer topic 1 question 251 discussion

Actual exam question from Google's Professional Machine Learning Engineer
Question #: 251
Topic #: 1
[All Professional Machine Learning Engineer Questions]

You are developing a training pipeline for a new XGBoost classification model based on tabular data. The data is stored in a BigQuery table. You need to complete the following steps:

1. Randomly split the data into training and evaluation datasets in a 65/35 ratio
2. Conduct feature engineering
3. Obtain metrics for the evaluation dataset
4. Compare models trained in different pipeline executions

How should you execute these steps?

  • A. 1. Using Vertex AI Pipelines, add a component to divide the data into training and evaluation sets, and add another component for feature engineering.
    2. Enable autologging of metrics in the training component.
    3. Compare pipeline runs in Vertex AI Experiments.
  • B. 1. Using Vertex AI Pipelines, add a component to divide the data into training and evaluation sets, and add another component for feature engineering.
    2. Enable autologging of metrics in the training component.
    3. Compare models using the artifacts’ lineage in Vertex ML Metadata.
  • C. 1. In BigQuery ML, use the CREATE MODEL statement with BOOSTED_TREE_CLASSIFIER as the model type and use BigQuery to handle the data splits.
    2. Use a SQL view to apply feature engineering and train the model using the data in that view.
    3. Compare the evaluation metrics of the models by using a SQL query with the ML.TRAINING_INFO statement.
  • D. 1. In BigQuery ML, use the CREATE MODEL statement with BOOSTED_TREE_CLASSIFIER as the model type and use BigQuery to handle the data splits.
    2. Use ML TRANSFORM to specify the feature engineering transformations and tram the model using the data in the table.
    3. Compare the evaluation metrics of the models by using a SQL query with the ML.TRAINING_INFO statement.
Show Suggested Answer Hide Answer
Suggested Answer: A 🗳️

Comments

Chosen Answer:
This is a voting comment (?) , you can switch to a simple comment.
Switch to a voting comment New
pikachu007
Highly Voted 10 months, 2 weeks ago
Selected Answer: A
Option B: While Vertex ML Metadata provides artifact lineage, it's less comprehensive for model comparison than Experiments. Options C and D: BigQuery ML is powerful for in-database model training, but it has limitations in pipeline orchestration, complex feature engineering, and detailed model comparison features, making it less suitable for this scenario.
upvoted 6 times
...
wences
Most Recent 2 months, 1 week ago
Selected Answer: A
Can anyone give a good reason for the answers without using ChatGPT or Gemini?
upvoted 1 times
...
tardigradum
3 months, 1 week ago
Selected Answer: A
BQ ML falls a bit short when it comes to building pipelines that include feature engineering and experiment comparison (it's better to use Vertex Pipelines and do the comparisons using Vertex Experiments).
upvoted 1 times
...
fitri001
7 months, 1 week ago
Selected Answer: A
Flexibility and Control: Vertex AI Pipelines allow you to define a custom pipeline with separate components for data splitting, feature engineering, and XGBoost training using your preferred libraries (like BigQueryClient and xgboost). This provides more control and customization compared to BigQuery ML's limited model types and functionality. Feature Engineering and Data Splitting: Separate components enable clear separation of concerns and potentially parallel execution for efficiency. Autologging and Model Comparison: Vertex AI autologging simplifies capturing evaluation metrics during training. Vertex AI Experiments offer a centralized interface to compare metrics across different pipeline runs (potentially with varying hyperparameter configurations).
upvoted 1 times
fitri001
7 months, 1 week ago
why not C & D? C & D. BigQuery ML: While BigQuery ML offers some XGBoost functionality, it has limitations: Limited Model Types: BigQuery ML doesn't provide the full flexibility of using custom XGBoost libraries with advanced configurations. Less Control over Feature Engineering: Feature engineering using SQL views might be restrictive compared to a dedicated component in Vertex AI Pipelines. Limited Model Comparison: While ML.TRAINING_INFO provides some insights, Vertex AI Experiments offer a more comprehensive view for comparing models across pipeline runs.
upvoted 1 times
...
...
pinimichele01
7 months, 2 weeks ago
Selected Answer: A
see b1a8fae
upvoted 1 times
...
omermahgoub
7 months, 2 weeks ago
Selected Answer: A
A: Leverage Vertex AI Pipelines and Experiments
upvoted 1 times
...
guilhermebutzke
9 months, 1 week ago
Selected Answer: A
My Answer: A A: CORRECT: It involves proper data splitting into training and evaluation sets and conducting feature engineering within the pipeline, fulfilling steps 1 and 2. Enabling autologging of metrics ensures that you can track and compare the performance of different model executions, fulfilling step 3. B: Not Correct: Better use Vertex AI Experiments C and D: Not Correct: BigQuery ML lacks functionalities for comparing models across pipeline runs. You would need to rely on external tools or custom scripts to extract and compare evaluation metrics, making the process less streamlined.
upvoted 2 times
...
b1a8fae
10 months, 1 week ago
Selected Answer: A
Compare models in different pipeline executions -> go for Vertex AI experiments
upvoted 3 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...