exam questions

Exam AWS Certified Machine Learning Engineer - Associate MLA-C01 All Questions

View all questions & answers for the AWS Certified Machine Learning Engineer - Associate MLA-C01 exam

Exam AWS Certified Machine Learning Engineer - Associate MLA-C01 topic 1 question 14 discussion

Case study -
An ML engineer is developing a fraud detection model on AWS. The training dataset includes transaction logs, customer profiles, and tables from an on-premises MySQL database. The transaction logs and customer profiles are stored in Amazon S3.
The dataset has a class imbalance that affects the learning of the model's algorithm. Additionally, many of the features have interdependencies. The algorithm is not capturing all the desired underlying patterns in the data.
The ML engineer needs to use an Amazon SageMaker built-in algorithm to train the model.
Which algorithm should the ML engineer use to meet this requirement?

  • A. LightGBM
  • B. Linear learner
  • C. К-means clustering
  • D. Neural Topic Model (NTM)
Show Suggested Answer Hide Answer
Suggested Answer: A 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
aragon_saa
Highly Voted 1 month, 2 weeks ago
Selected Answer: B
Answer is B
upvoted 6 times
...
Leo2023aws
Highly Voted 1 month, 2 weeks ago
Selected Answer: A
https://docs.aws.amazon.com/en_kr/sagemaker/latest/dg/lightgbm.html
upvoted 5 times
...
xukun
Most Recent 1 day, 7 hours ago
Selected Answer: A
https://docs.aws.amazon.com/en_kr/sagemaker/latest/dg/lightgbm.html
upvoted 1 times
...
Makendran
1 week ago
Selected Answer: B
In an ideal scenario, for a problem with these characteristics (fraud detection, class imbalance, feature interdependencies, complex patterns), a tree-based ensemble method like XGBoost (which is a SageMaker built-in algorithm) would be more suitable. XGBoost can handle non-linear relationships, is robust to class imbalance with proper tuning, and can capture complex patterns in the data. However, given the options provided and the requirement to use a SageMaker built-in algorithm, the Linear learner is the best available choice among these options for this specific fraud detection task.
upvoted 1 times
...
gulf1324
1 week, 3 days ago
Selected Answer: B
A. Light BGM : It's suitable model, but not built-in model for SageMaker. Answer B. Linear learner : suitable model, built-in model for SageMaker. C. K-means clustering : groups similar data points, not suitable for classification problems, and it's unsupervised learning algorithm so doesn't fit in this case(fraud detection). D. Neural Topic Model: used for topic modeling and document classification, not suitable for fraud detection
upvoted 2 times
minhhnh
6 days, 1 hour ago
Light BGM is built-in model for SageMaker https://docs.aws.amazon.com/sagemaker/latest/dg/lightgbm.html
upvoted 1 times
...
...
khchan123
1 week, 5 days ago
Selected Answer: A
Here's why LightGBM is the most suitable algorithm for this fraud detection task: Handling Class Imbalance: LightGBM is particularly effective at handling imbalanced datasets, which is a key issue mentioned in the problem statement. It has built-in mechanisms to deal with class imbalance. Feature Interdependencies: LightGBM can capture complex feature interactions through its tree-based structure, addressing the issue of feature interdependencies mentioned in the problem. Capturing Underlying Patterns: As an advanced gradient boosting framework, LightGBM is excellent at capturing complex patterns in data, which the current algorithm is struggling with. Suitable for Fraud Detection: LightGBM is widely used in fraud detection tasks due to its high performance and ability to handle large datasets efficiently. Handling Various Data Types: It can work well with the mix of data types likely present in transaction logs, customer profiles, and database tables.
upvoted 1 times
...
ninomfr64
1 week, 6 days ago
Selected Answer: A
We have an unbalanced dataset, this means we have labelled dataset thus we are going to use a supervised model training. This reduce options to A and B (K-means and NTM are unsupervised). Both LightGBM and Linear Learner provides hyperparameter to manage unbalanced datasets, respectively "scale-pos_weight" and "positive_example_weight_mult". I would go for LightGBM as this algorithm is more suited to handle complex relationship among features, while Linear Learner learns a linear function, or, for classification problems, a linear threshold function, and maps a vector x to an approximation of the label y.
upvoted 1 times
...
Ell89
2 weeks ago
Selected Answer: B
Linear Learner. LightGBM is NOT a built in algorithm which the question asks for.
upvoted 1 times
...
michaelcloud
2 weeks, 2 days ago
Selected Answer: A
This is a binary classification problem so LightGBM so be used. Other algorithms are not for binary classification.
upvoted 1 times
...
bakju0
2 weeks, 2 days ago
Selected Answer: B
Is LightGBM Built-in? Technically, no, LightGBM is not a “built-in algorithm” in the same category as SageMaker’s core algorithms (like Linear Learner or XGBoost). However, it is supported through prebuilt containers, which makes it easy to use in SageMaker.
upvoted 2 times
...
Saransundar
1 month, 1 week ago
Selected Answer: A
A. LightGBM: Handles class imbalance; captures feature interdependencies; models complex patterns. B. Linear Learner: Limited with interdependent features; struggles with complex patterns; suitable for linear relationships. C. K-means Clustering: Unsupervised algorithm; not suitable for classification; can't handle class imbalance. D. Neural Topic Model (NTM): Designed for topic modeling; unsuitable for fraud detection; doesn't address class imbalance.
upvoted 2 times
...
Linux_master
1 month, 2 weeks ago
Selected Answer: A
This is a binary classification problem so LightGBM so be used. Other algorithms are not for binary classification.
upvoted 3 times
...
GiorgioGss
1 month, 2 weeks ago
Selected Answer: A
Light Gradient Boosting Machine is effective for handling class imbalances and feature interdependencies.
upvoted 1 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago