Welcome to ExamTopics
ExamTopics Logo
- Expert Verified, Online, Free.
exam questions

Exam Professional Machine Learning Engineer All Questions

View all questions & answers for the Professional Machine Learning Engineer exam

Exam Professional Machine Learning Engineer topic 1 question 269 discussion

Actual exam question from Google's Professional Machine Learning Engineer
Question #: 269
Topic #: 1
[All Professional Machine Learning Engineer Questions]

You are developing a model to help your company create more targeted online advertising campaigns. You need to create a dataset that you will use to train the model. You want to avoid creating or reinforcing unfair bias in the model. What should you do? (Choose two.)

  • A. Include a comprehensive set of demographic features
  • B. Include only the demographic groups that most frequently interact with advertisements
  • C. Collect a random sample of production traffic to build the training dataset
  • D. Collect a stratified sample of production traffic to build the training dataset
  • E. Conduct fairness tests across sensitive categories and demographics on the trained model
Show Suggested Answer Hide Answer
Suggested Answer: D 🗳️

Comments

Chosen Answer:
This is a voting comment (?) , you can switch to a simple comment.
Switch to a voting comment New
wences
2 months, 1 week ago
Selected Answer: D
From my statistical point of view, D and E will mitigate the effect of bias.
upvoted 1 times
...
AzureDP900
4 months, 3 weeks ago
D and E is right answer, question asks us to select 2 right answers • To avoid creating or reinforcing unfair bias in the model, you should collect a representative and diverse dataset (option D) that includes a stratified sample of production traffic. This ensures that your training data is inclusive and accurately represents the diversity of your target audience. • Once you have collected your training dataset, you should conduct fairness tests across sensitive categories and demographics on the trained model (option E). This involves evaluating whether the model treats different demographic groups fairly and without bias. If biases are detected, you can take steps to mitigate them and ensure that your model is fair and accurate.
upvoted 1 times
...
AzureDP900
4 months, 3 weeks ago
D and E is right answer, question asks us to select 2 right answers • To avoid creating or reinforcing unfair bias in the model, you should collect a representative and diverse dataset (option D) that includes a stratified sample of production traffic. This ensures that your training data is inclusive and accurately represents the diversity of your target audience. • Once you have collected your training dataset, you should conduct fairness tests across sensitive categories and demographics on the trained model (option E). This involves evaluating whether the model treats different demographic groups fairly and without bias. If biases are detected, you can take steps to mitigate them and ensure that your model is fair and accurate.
upvoted 1 times
...
dija123
4 months, 3 weeks ago
Selected Answer: D
Agree with D and E
upvoted 1 times
...
omermahgoub
7 months, 2 weeks ago
Selected Answer: D
D. Stratified sampling to ensure the different demographic groups or categories are proportionally represented in the training data. This helps mitigate bias that might arise if certain groups are under-represented. E. Fairness tests can reveal disparities in how the model treats different populations, allowing you to identify and address potential biases.
upvoted 3 times
...
MultiCloudIronMan
7 months, 3 weeks ago
Selected Answer: D
D and E is the two answers. Two selections are required
upvoted 2 times
pinimichele01
7 months, 2 weeks ago
why not D and A?
upvoted 2 times
...
...
CHARLIE2108
9 months, 1 week ago
Selected Answer: D
I went D, E
upvoted 2 times
...
guilhermebutzke
9 months, 1 week ago
Selected Answer: D
DE D. Collect a stratified sample of production traffic to build the training dataset: This ensures that the training data represents the diverse demographics that will be targeted by the advertising campaigns. Random sampling might unintentionally underrepresent certain groups, leading to biased model outputs. E. Conduct fairness tests across sensitive categories and demographics on the trained model: This allows you to identify and address any potential biases that may have emerged during the training process. Evaluating the model's performance on different groups helps ensure fair and responsible deployment.
upvoted 1 times
...
daidai75
10 months, 1 week ago
Selected Answer: D
I go for D & E: A stratified sample ensures that the training data represents the distribution of the target population across relevant demographics or other sensitive categories. This helps mitigate bias arising from underrepresented groups in the data. Regularly testing the model for fairness across sensitive categories helps identify and address potential bias issues before deploying the model in production. This can involve metrics like precision, recall, and F1 score for different demographic groups.
upvoted 1 times
...
b1a8fae
10 months, 1 week ago
Selected Answer: D
D E. ChatGPT explanation below (but I think makes quite a lot of sense) Collect a Stratified Sample (Option D): Stratified sampling involves dividing the population into subgroups (strata) and then randomly sampling from each subgroup. This ensures that the training dataset represents the diversity of the population, helping to avoid biases. By collecting a stratified sample of production traffic, you are more likely to have a balanced representation of different demographic groups, reducing the risk of biased model outcomes. Conduct Fairness Tests (Option E): After training the model, it's crucial to conduct fairness tests to evaluate its performance across different sensitive categories and demographics. This involves measuring the model's predictions and outcomes for various groups to identify any disparities. Fairness tests help you assess and address biases that may have been inadvertently introduced during the training process.
upvoted 3 times
...
shadz10
10 months, 1 week ago
Selected Answer: C
C, D - Conducting fairness tests across sensitive categories and demographics on the trained model is indeed important. However, this option focuses on post-training analysis rather than dataset creation. While it's a crucial step for ensuring fairness, it doesn't directly address how to create a training dataset to avoid bias. Hence C,D
upvoted 1 times
tavva_prudhvi
9 months, 2 weeks ago
Check b1a8fae comment on why D is better than C!
upvoted 1 times
...
...
pikachu007
10 months, 2 weeks ago
Selected Answer: D
D. Stratified Sampling: Randomly sampling your data might not accurately represent the diversity of your target audience, potentially introducing bias by over- or under-representing certain demographics. Stratified sampling ensures your training dataset reflects the distribution of sensitive features (e.g., age, gender, income) observed in your production traffic, helping mitigate bias during model training. E. Fairness Testing: Simply collecting unbiased data isn't enough. Regularly testing your trained model for fairness across sensitive categories is crucial. This involves measuring and analyzing metrics like accuracy, precision, recall, and F1 score for different demographic groups. Identifying disparities in performance can trigger further investigation and potential re-training to address bias.
upvoted 1 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...