exam questions

Exam Professional Machine Learning Engineer All Questions

View all questions & answers for the Professional Machine Learning Engineer exam

Exam Professional Machine Learning Engineer topic 1 question 255 discussion

Actual exam question from Google's Professional Machine Learning Engineer
Question #: 255
Topic #: 1
[All Professional Machine Learning Engineer Questions]

You have recently used TensorFlow to train a classification model on tabular data. You have created a Dataflow pipeline that can transform several terabytes of data into training or prediction datasets consisting of TFRecords. You now need to productionize the model, and you want the predictions to be automatically uploaded to a BigQuery table on a weekly schedule. What should you do?

  • A. Import the model into Vertex AI and deploy it to a Vertex AI endpoint. On Vertex AI Pipelines, create a pipeline that uses the DataflowPythonJobOp and the ModelBacthPredictOp components.
  • B. Import the model into Vertex AI and deploy it to a Vertex AI endpoint. Create a Dataflow pipeline that reuses the data processing logic sends requests to the endpoint, and then uploads predictions to a BigQuery table.
  • C. Import the model into Vertex AI. On Vertex AI Pipelines, create a pipeline that uses the
    DataflowPvthonJobOp and the ModelBatchPredictOp components.
  • D. Import the model into BigQuery. Implement the data processing logic in a SQL query. On Vertex AI Pipelines create a pipeline that uses the BigquervQueryJobOp and the BigqueryPredictModelJobOp components.
Show Suggested Answer Hide Answer
Suggested Answer: C 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
BlehMaks
Highly Voted 1 year, 2 months ago
Selected Answer: C
The DataflowPythonJobOp operator lets you create a Vertex AI Pipelines component that prepares data by submitting a Python-based Apache Beam job to Dataflow for execution. https://cloud.google.com/vertex-ai/docs/pipelines/dataflow-component#dataflowpythonjobop Using we can specify an output location for Vertex AI to store predictions results https://cloud.google.com/vertex-ai/docs/pipelines/batchprediction-component A - is incorrect since we dont need an endpoint for batch predictions B - creating a new Dataflow pipeline is redundant
upvoted 10 times
...
lunalongo
Most Recent 4 months, 1 week ago
Selected Answer: C
C is the best option because it uses: 1) Vertex AI Pipelines for orchestrating the flow (managed and scalable). 2) DataflowPythonJobOp for prep and ModelBatchPredictOp for batch predictions on Vertex AI. *A deploys the model to a Vertex AI endpoint, inefficient for batch jobs! *B uses a single Dataflow pipeline, which needs custom Vertex AI and BQ integration. *D uses BigQuery, a datawarehouse, for model deployment and prediction.
upvoted 1 times
...
AK2020
8 months ago
Selected Answer: B
Uploading predictions directly to BigQuery from the Dataflow pipeline integrates seamlessly with your data storage.
upvoted 1 times
...
AzureDP900
9 months ago
B is right because 1)You've already trained a classification model using TensorFlow, so you need to productionize it by deploying it to a Vertex AI endpoint. 2)To automate the prediction process on a weekly schedule, you can create a Dataflow pipeline that reuses your existing data processing logic. This pipeline will send requests to the deployed model for inference and then upload the predicted results to BigQuery.
upvoted 1 times
...
Prakzz
9 months ago
Selected Answer: B
Only option B talks about loading the data to BigQuery
upvoted 1 times
...
rcapj
9 months, 2 weeks ago
B Vertex AI Deployment: Vertex AI provides a managed environment for deploying machine learning models. It simplifies the process and ensures scalability. Dataflow Pipeline Reuse: Reusing the existing Dataflow pipeline for data processing leverages your existing code and avoids redundant logic. Model Endpoint Predictions: Sending requests to the deployed model endpoint allows for efficient prediction generation. BigQuery Upload: Uploading predictions directly to BigQuery from the Dataflow pipeline integrates seamlessly with your data storage.
upvoted 2 times
...
gscharly
11 months, 2 weeks ago
Selected Answer: C
No need to deploy to endpoint as we need batch predictions. ModelBatchPredictOp can upload data to BQ. Dataflow pipeline logic can be implemented in DataflowPythonJobOp
upvoted 4 times
...
fitri001
11 months, 3 weeks ago
Selected Answer: B
TFRecords is a specific file format designed by TensorFlow for storing data in a way that's efficient for the machine learning framework. Here are some key points about TFRecords:
upvoted 1 times
...
fitri001
11 months, 3 weeks ago
Selected Answer: B
Option A: Vertex AI Pipelines' ModelBatchPredictOp is designed for batch prediction within pipelines, not for serving models through an endpoint. Option C: Importing the model directly into BigQuery is not feasible for TensorFlow models. Option D: Vertex AI Pipelines' BigqueryPredictModelJobOp assumes the model is already trained and hosted in BigQuery ML, which isn't the case here.
upvoted 2 times
pinimichele01
11 months, 3 weeks ago
Importing the model directly into BigQuery is not feasible for TensorFlow models. -> not true
upvoted 3 times
...
...
pinimichele01
11 months, 3 weeks ago
Selected Answer: C
ModelBatchPredictOp -> upload automatically on BQ No need for endpoint --> C
upvoted 2 times
...
pinimichele01
12 months ago
Selected Answer: C
agree with BlehMaks
upvoted 1 times
...
pertoise
1 year, 1 month ago
Answer is C. No need for an endpoint here : Simply specify the BigQuery table URI in the ModelBatchPredictOp parameter and you're done automatically uploading to BigQuery
upvoted 3 times
...
guilhermebutzke
1 year, 1 month ago
Selected Answer: B
My Answer: B The most complete answer, and reuse a created pipeline. Don’t make sense to use DataflowPythonJobOp when you have already created a dataflow pipeline that does the same.
upvoted 2 times
...
tavva_prudhvi
1 year, 1 month ago
Selected Answer: B
Not A, C as they does not explicitly mention how the predictions will be uploaded to BigQuery.
upvoted 1 times
...
daidai75
1 year, 2 months ago
Selected Answer: B
The answer is B, optional A and B doesn't mention how to import prediction result to BigQuery.
upvoted 1 times
...
pikachu007
1 year, 2 months ago
Selected Answer: B
Option A: Vertex AI Pipelines are excellent for orchestrating ML workflows but might not be as efficient as Dataflow for large-scale data processing, especially with existing Dataflow logic. Option C: While Vertex AI Pipelines can handle model loading and prediction, Dataflow is better suited for large-scale data processing and BigQuery integration. Option D: BigQuery ML is primarily for in-database model training and prediction, not ideal for external models or large-scale data processing.
upvoted 1 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago