exam questions

Exam Professional Machine Learning Engineer All Questions

View all questions & answers for the Professional Machine Learning Engineer exam

Exam Professional Machine Learning Engineer topic 1 question 265 discussion

Actual exam question from Google's Professional Machine Learning Engineer
Question #: 265
Topic #: 1
[All Professional Machine Learning Engineer Questions]

You work for a company that is developing an application to help users with meal planning. You want to use machine learning to scan a corpus of recipes and extract each ingredient (e.g., carrot, rice, pasta) and each kitchen cookware (e.g., bowl, pot, spoon) mentioned. Each recipe is saved in an unstructured text file. What should you do?

  • A. Create a text dataset on Vertex AI for entity extraction Create two entities called “ingredient” and “cookware”, and label at least 200 examples of each entity. Train an AutoML entity extraction model to extract occurrences of these entity types. Evaluate performance on a holdout dataset.
  • B. Create a multi-label text classification dataset on Vertex AI. Create a test dataset, and label each recipe that corresponds to its ingredients and cookware. Train a multi-class classification model. Evaluate the model’s performance on a holdout dataset.
  • C. Use the Entity Analysis method of the Natural Language API to extract the ingredients and cookware from each recipe. Evaluate the model's performance on a prelabeled dataset.
  • D. Create a text dataset on Vertex AI for entity extraction. Create as many entities as there are different ingredients and cookware. Train an AutoML entity extraction model to extract those entities. Evaluate the model’s performance on a holdout dataset.
Show Suggested Answer Hide Answer
Suggested Answer: A 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
Wuthuong1234
1 month ago
Selected Answer: C
The Entity detection in the NLP API will be sufficient to identify ingredients and cookware-related words. It is much easier than training your own model in AutoML. Keep in mind that training on your own dataset could introduce some bias. Imagine your training data might cover many French or western recipes, but suddenly you get lots of Thai recipes in production. Your AutoML model would struggle to correctly identify ingredients that are not so common in western cooking such as lemongrass, kecap manis, kaffir or galangal.
upvoted 1 times
...
andrea_c_
3 months, 2 weeks ago
Selected Answer: C
With A you must label a dataset. Since the entities that need to be recognized are pretty common this effort is not justified. Moreover, as specified in https://cloud.google.com/vertex-ai/docs/text-data/entity-extraction/prepare-data, "You must supply at least 1, and no more than 100, unique labels to annotate entities that you want to extract." So, it looks like the dataset has a limit of 100 entities, which I do not think is enough for this use case.
upvoted 1 times
...
Omi_04040
3 months, 3 weeks ago
Selected Answer: A
This option involves creating a dataset specifically for entity extraction and training an AutoML model to identify ingredients and cookware. By labeling a minimum of 200 instances for each entity, it ensures a sufficient amount of data for training. Using a holdout dataset for assessment helps evaluate the model's performance. Overall, this approach seems appropriate for the task at hand. Reference: https://cloud.google.com/vertex-ai/docs/text-data/entity-extraction/prepare-data
upvoted 1 times
...
AzureDP900
9 months ago
By choosing option A, you can leverage the power of machine learning to efficiently extract ingredients and cookware from recipes in a scalable manner. option C uses the Entity Analysis method of the Natural Language API, which might be a viable option if you had access to the API's pre-trained models. However, since you're working with Vertex AI, creating a dataset for entity extraction is a better choice.
upvoted 1 times
...
fitri001
11 months, 3 weeks ago
Selected Answer: A
For extracting ingredients and cookware from recipe text files, creating a text dataset on Vertex AI for entity extraction with a custom NER model is the better approach. While it requires more upfront effort for data labeling and training, it offers superior accuracy and control over the types of entities extracted. However, if you need a quick and easy solution to get started, the Natural Language API's Entity Analysis can be a temporary option. Be aware that the accuracy might be lower, and you might need to post-process the results to filter out irrelevant entities.
upvoted 2 times
...
omermahgoub
11 months, 3 weeks ago
Selected Answer: C
Natural Language API offers a pre-built solution for entity analysis which eliminates the need for custom model training and labeling large datasets, saving time and resources. Vertex AI AutoML can aslo be used for entity extraction but it requires data labeling and training, which can be time-consuming for a vast number of potential ingredients and cookware.
upvoted 1 times
...
guilhermebutzke
1 year, 1 month ago
Selected Answer: A
My Answer: A  A: is the most suitable approach for this task because we need to identify and extract specific named entities ("ingredient" and "cookware") from the text, not classify the entire recipe into predefined categories. B: This approach would require classifying each recipe based on all possible ingredients and cookware, leading to a vast number of classes and potential performance issues. C: This pre-built solution might not be as customizable or scalable as training a specific model for this task. D: This is impractical and unnecessary as the number of potential ingredients and cookware is vast.
upvoted 2 times
...
daidai75
1 year, 2 months ago
I prefer to A. Option C is not the best, because The NLP API is designed to identify general entities within text. While it's effective for broad categories, it may not be as precise for specialized domains like cooking ingredients and cookware, which require a more tailored approach.
upvoted 2 times
...
b1a8fae
1 year, 2 months ago
Selected Answer: A
A. "... you might create an entity extraction model to identify specialized terminology in legal documents or patents." I prefer this over C, which might classify carrot as vegetable, chicken as meat... custom entity extraction allows you to specify what entities you wish to extract from the text.
upvoted 4 times
b1a8fae
1 year, 2 months ago
https://cloud.google.com/vertex-ai/docs/text-data/entity-extraction/prepare-data
upvoted 3 times
...
...
shadz10
1 year, 2 months ago
Selected Answer: C
Reconsidering my answer and going with C Option A involves using AutoML entity extraction, which could be a valid approach. However, for extracting entities like ingredients and cookware, Google Cloud's pre-trained Natural Language API might be a more straightforward solution.
upvoted 1 times
tavva_prudhvi
1 year, 1 month ago
No, A is right as it may not be as effective for this specific task unless the ingredients and cookware are already well-represented within the types of entities the API is trained to recognize. This approach might require less initial setup but could be less accurate for specialized domains like recipes.
upvoted 1 times
...
...
shadz10
1 year, 2 months ago
Selected Answer: A
A is the correct option here
upvoted 2 times
...
pikachu007
1 year, 2 months ago
Selected Answer: C
Option B: Multi-label text classification is less suitable for identifying specific entities within text and would require labeling entire recipes with multiple classes, increasing complexity and reducing model specificity. Option C: Natural Language API's Entity Analysis might not be as accurate for this specialized domain as a model trained on custom recipe data. Option D: Creating separate entities for each ingredient and cookware type would significantly increase labeling effort and potentially hinder model generalization.
upvoted 1 times
kalle_balle
1 year, 2 months ago
do you mean Option A?
upvoted 3 times
...
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago