Welcome to ExamTopics
ExamTopics Logo
- Expert Verified, Online, Free.
exam questions

Exam Professional Machine Learning Engineer All Questions

View all questions & answers for the Professional Machine Learning Engineer exam

Exam Professional Machine Learning Engineer topic 1 question 165 discussion

Actual exam question from Google's Professional Machine Learning Engineer
Question #: 165
Topic #: 1
[All Professional Machine Learning Engineer Questions]

You want to train an AutoML model to predict house prices by using a small public dataset stored in BigQuery. You need to prepare the data and want to use the simplest, most efficient approach. What should you do?

  • A. Write a query that preprocesses the data by using BigQuery and creates a new table. Create a Vertex AI managed dataset with the new table as the data source.
  • B. Use Dataflow to preprocess the data. Write the output in TFRecord format to a Cloud Storage bucket.
  • C. Write a query that preprocesses the data by using BigQuery. Export the query results as CSV files, and use those files to create a Vertex AI managed dataset.
  • D. Use a Vertex AI Workbench notebook instance to preprocess the data by using the pandas library. Export the data as CSV files, and use those files to create a Vertex AI managed dataset.
Show Suggested Answer Hide Answer
Suggested Answer: A 🗳️

Comments

Chosen Answer:
This is a voting comment (?) , you can switch to a simple comment.
Switch to a voting comment New
PhilipKoku
5 months, 2 weeks ago
Selected Answer: A
A) Keep the data in BigQuery and create a new table to avoid latency moving data out of BigQuery
upvoted 2 times
...
nmnm22
5 months, 3 weeks ago
Selected Answer: A
A seems the correct one
upvoted 1 times
...
gscharly
7 months, 1 week ago
Selected Answer: A
I go for A:
upvoted 1 times
...
shadz10
10 months, 1 week ago
Selected Answer: A
can export directly from big query as vertex ai managed dataset to use train an autoML model
upvoted 1 times
...
36bdc1e
10 months, 2 weeks ago
A By writing a query that preprocesses the data using BigQuery and creating a new table, you can directly create a Vertex AI managed dataset with the new table as the data source. This approach is efficient because it leverages BigQuery’s powerful data processing capabilities and avoids the need to export data to another format or service. It also simplifies the process by keeping everything within the Google Cloud ecosystem. This makes it easier to manage and monitor your data and model training process.
upvoted 2 times
...
vale_76_na_xxx
10 months, 2 weeks ago
I go for A:
upvoted 2 times
...
b1a8fae
10 months, 2 weeks ago
Selected Answer: A
Forgot to vote
upvoted 1 times
...
b1a8fae
10 months, 2 weeks ago
A seems the easiest to me: preprocess the data on BigQuery (where the input table is stored) and export directly as Vertex AI managed dataset.
upvoted 2 times
...
kalle_balle
10 months, 2 weeks ago
Selected Answer: B
Dataflow seems like the easiest and most scalable way to deal with this issue. Option B.
upvoted 1 times
f084277
1 week, 1 day ago
The data is already in BigQuery. Preprocess the data in BigQuery. How is Dataflow easier than BigQuery? (question doesn't mention anything about scalability)
upvoted 1 times
...
pinimichele01
6 months, 4 weeks ago
small dataset -> no dataflow
upvoted 1 times
...
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...