Welcome to ExamTopics
ExamTopics Logo
- Expert Verified, Online, Free.
exam questions

Exam AWS Certified Machine Learning - Specialty All Questions

View all questions & answers for the AWS Certified Machine Learning - Specialty exam

Exam AWS Certified Machine Learning - Specialty topic 1 question 123 discussion

A Data Scientist is developing a machine learning model to classify whether a financial transaction is fraudulent. The labeled data available for training consists of
100,000 non-fraudulent observations and 1,000 fraudulent observations.
The Data Scientist applies the XGBoost algorithm to the data, resulting in the following confusion matrix when the trained model is applied to a previously unseen validation dataset. The accuracy of the model is 99.1%, but the Data Scientist needs to reduce the number of false negatives.

Which combination of steps should the Data Scientist take to reduce the number of false negative predictions by the model? (Choose two.)

  • A. Change the XGBoost eval_metric parameter to optimize based on Root Mean Square Error (RMSE).
  • B. Increase the XGBoost scale_pos_weight parameter to adjust the balance of positive and negative weights.
  • C. Increase the XGBoost max_depth parameter because the model is currently underfitting the data.
  • D. Change the XGBoost eval_metric parameter to optimize based on Area Under the ROC Curve (AUC).
  • E. Decrease the XGBoost max_depth parameter because the model is currently overfitting the data.
Show Suggested Answer Hide Answer
Suggested Answer: BD 🗳️

Comments

Chosen Answer:
This is a voting comment (?) , you can switch to a simple comment.
Switch to a voting comment New
LydiaGom
Highly Voted 2 years, 5 months ago
B and D
upvoted 12 times
...
ovokpus
Highly Voted 2 years, 3 months ago
Selected Answer: BD
Compensate for imbalance and optimize on AUC. This is a class imbalance problem, not an overfitting problem.
upvoted 8 times
rb39
2 years, 1 month ago
totally right, overfitting has nothing to do so there is no need to reduce tree depth
upvoted 1 times
...
...
MJSY
Most Recent 2 weeks, 1 day ago
Selected Answer: BD
the question didnt show the model performance on training data, so the overfitting issues is not correct.
upvoted 1 times
...
loict
1 year, 1 month ago
Selected Answer: BD
A. NO - that will not address FN specifically but also FP B. YES - changing weight is best practice for class imbalance C. NO - there is no underfitting at 99.1% accuracy D. YES - AUC will address recall, which takes into account FN rate E. NO - there is no overfitting at 99.1% accuracy
upvoted 3 times
...
Mickey321
1 year, 1 month ago
Selected Answer: BD
Step B: Increasing the XGBoost scale_pos_weight parameter to adjust the balance of positive and negative weights can help the model deal with the imbalanced dataset. According to the XGBoost documentation, this parameter controls the balance of positive and negative weights, and is useful for unbalanced classes. A typical value to consider is sum(negative instances) / sum(positive instances). In this case, since there are 100 times more non-fraudulent transactions than fraudulent ones, setting scale_pos_weight to 100 can make the model more sensitive to the minority class and reduce false negatives. Step D: Changing the XGBoost eval_metric parameter to optimize based on Area Under the ROC Curve (AUC) can help the model focus on improving the true positive rate and the true negative rate, which are both important for fraud detection. According to the XGBoost
upvoted 2 times
...
Mickey321
1 year, 1 month ago
Selected Answer: BD
Step B: Increasing the XGBoost scale_pos_weight parameter to adjust the balance of positive and negative weights can help the model deal with the imbalanced dataset. According to the XGBoost documentation, this parameter controls the balance of positive and negative weights, and is useful for unbalanced classes. A typical value to consider is sum(negative instances) / sum(positive instances). In this case, since there are 100 times more non-fraudulent transactions than fraudulent ones, setting scale_pos_weight to 100 can make the model more sensitive to the minority class and reduce false negatives. Step D: Changing the XGBoost eval_metric parameter to optimize based on Area Under the ROC Curve (AUC) can help the model focus on improving the true positive rate and the true negative rate, which are both important for fraud detection.
upvoted 1 times
...
Mllb
1 year, 6 months ago
Selected Answer: BE
I have some doubts about D and E. Precision-Recall AUC is better than AUC curve in imbalanced classes. Then, I choose E
upvoted 2 times
...
AjoseO
1 year, 7 months ago
Selected Answer: BD
Option A and Option E are unlikely to help reduce false negatives. Option C, increasing max_depth, may lead to overfitting, which could make the model worse. Option D, changing the eval_metric to optimize based on AUC, could help improve the model's ability to discriminate between the two classes. Option B, increasing the scale_pos_weight parameter to adjust the balance of positive and negative weights, can help the model better handle imbalanced datasets, which is the case here. By increasing the weight of positive examples, the model will learn to prioritize correctly classifying them, which should reduce the number of false negatives.
upvoted 1 times
...
Tomatoteacher
1 year, 9 months ago
Selected Answer: BD
BD, I have done this before, but it would be better to use Average Precision(AP) instead of AUC, but it is better than other answers.
upvoted 1 times
...
Shailendraa
2 years, 1 month ago
12-sep exam
upvoted 1 times
...
ovokpus
2 years, 3 months ago
Selected Answer: BE
Compensate for imbalance and overwriting.
upvoted 1 times
...
NeverMinda
2 years, 4 months ago
Selected Answer: BE
B and E
upvoted 1 times
...
MLGuru
2 years, 5 months ago
B. Increase the XGBoost scale_pos_weight parameter to adjust the balance of positive and negative weights is the correct answer. According to https://docs.aws.amazon.com/sagemaker/latest/dg/xgboost_hyperparameters.html, scale_pos_weight controls the balance of positive and negative weights. It's useful for unbalanced classes.
upvoted 3 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...