New Year Sale 2026! Hurry Up, Grab the Special Discount - Save 25% - Ends In 00:00:00 Coupon code: SAVE25
Welcome to Pass4Success

- Free Preparation Discussions

Amazon MLS-C01 Exam - Topic 1 Question 107 Discussion

Actual exam question for Amazon's MLS-C01 exam
Question #: 107
Topic #: 1
[All MLS-C01 Questions]

An obtain relator collects the following data on customer orders: demographics, behaviors, location, shipment progress, and delivery time. A data scientist joins all the collected datasets. The result is a single dataset that includes 980 variables.

The data scientist must develop a machine learning (ML) model to identify groups of customers who are likely to respond to a marketing campaign.

Which combination of algorithms should the data scientist use to meet this requirement? (Select TWO.)

Show Suggested Answer Hide Answer
Suggested Answer: B

The best solution to meet the requirements is to tune the csv_weight hyperparameter and the scale_pos_weight hyperparameter by using automatic model tuning (AMT). Optimize on {''HyperParameterTuningJobObjective'': {''MetricName'': ''validation:f1'', ''Type'': ''Maximize''}}.

The csv_weight hyperparameter is used to specify the instance weights for the training data in CSV format. This can help handle imbalanced data by assigning higher weights to the minority class examples and lower weights to the majority class examples. The scale_pos_weight hyperparameter is used to control the balance of positive and negative weights. It is the ratio of the number of negative class examples to the number of positive class examples. Setting a higher value for this hyperparameter can increase the importance of the positive class and improve the recall. Both of these hyperparameters can help the XGBoost model capture as many instances of returned items as possible.

Automatic model tuning (AMT) is a feature of Amazon SageMaker that automates the process of finding the best hyperparameter values for a machine learning model. AMT uses Bayesian optimization to search the hyperparameter space and evaluate the model performance based on a predefined objective metric. The objective metric is the metric that AMT tries to optimize by adjusting the hyperparameter values. For imbalanced classification problems, accuracy is not a good objective metric, as it can be misleading and biased towards the majority class. A better objective metric is the F1 score, which is the harmonic mean of precision and recall. The F1 score can reflect the balance between precision and recall and is more suitable for imbalanced data. The F1 score ranges from 0 to 1, where 1 is the best possible value. Therefore, the type of the objective should be ''Maximize'' to achieve the highest F1 score.

By tuning the csv_weight and scale_pos_weight hyperparameters and optimizing on the F1 score, the data scientist can meet the requirements most cost-effectively. This solution requires tuning only two hyperparameters, which can reduce the computation time and cost compared to tuning all possible hyperparameters. This solution also uses the appropriate objective metric for imbalanced classification, which can improve the model performance and capture more instances of returned items.

References:

* XGBoost Hyperparameters

* Automatic Model Tuning

* How to Configure XGBoost for Imbalanced Classification

* Imbalanced Data


Contribute your Thoughts:

0/2000 characters
Layla
3 months ago
Factorization machines? Never heard of that one!
upvoted 0 times
...
Mireya
3 months ago
Wait, LDA for this? Seems off to me.
upvoted 0 times
...
Nada
3 months ago
PCA could help with dimensionality reduction too.
upvoted 0 times
...
Dorothy
4 months ago
Totally agree, K-means is a solid choice.
upvoted 0 times
...
Daniela
4 months ago
K-means is great for clustering!
upvoted 0 times
...
Freeman
4 months ago
I vaguely recall something about LDA being used for topic modeling, but I'm not sure how it applies to customer segmentation. It feels a bit off for this question.
upvoted 0 times
...
Freeman
4 months ago
I practiced a similar question where K-means was paired with another clustering algorithm. I feel like it could be a good fit for this scenario too.
upvoted 0 times
...
Sueann
4 months ago
I think PCA is more about dimensionality reduction, but it might help in preprocessing the data before clustering. I'm not entirely confident about that though.
upvoted 0 times
...
Lanie
5 months ago
I remember that K-means is often used for clustering, which seems relevant for identifying customer groups. But I'm not sure if it's the best choice here.
upvoted 0 times
...
Marci
5 months ago
This is a lot of data to work with! I'd definitely lean on some powerful algorithms like factorization machines to uncover the hidden patterns. Combine that with K-means clustering and I think I can crack this problem.
upvoted 0 times
...
Stephaine
5 months ago
Okay, I've got a strategy here. I'd use PCA to reduce the dimensionality, then apply K-means clustering to find the customer groups. That should give me the insights I need to identify the best targets for the marketing campaign.
upvoted 0 times
...
Carin
5 months ago
Hmm, I'm a bit confused by the "semantic feg mentation" option. That doesn't sound like a real algorithm to me. I think I'd stick to the more standard choices like K-means and LDA.
upvoted 0 times
...
Jeffrey
5 months ago
This is a tricky one! With 980 variables, I'd definitely want to use some dimensionality reduction techniques like PCA to start. Then I'd probably try a clustering algorithm like K-means to identify the customer groups.
upvoted 0 times
...
Elena
9 months ago
Latent Dirichlet Allocation? Sounds like a fancy way of saying 'we have no idea what's going on here'. I'll take K-means and PCA - at least those algorithms make sense, unlike 'Semantic fegmentation'. Did the question writer get their keyboard stuck or something?
upvoted 0 times
...
Crista
9 months ago
I'd say K-means and FM. Gotta love those Factorization Machines - they can handle all that juicy data! Although, I'm a bit worried about the 'Semantic fegmentation' option. Sounds like someone's been hitting the eggnog a little too hard.
upvoted 0 times
...
Lucy
9 months ago
K-means and PCA, for sure. But can we talk about the name 'Latent Dirichlet Allocation'? Sounds like a spell from Harry Potter!
upvoted 0 times
Yvonne
8 months ago
It does sound mysterious, but it's actually a topic modeling algorithm used in natural language processing.
upvoted 0 times
...
Jolene
8 months ago
Haha, 'Latent Dirichlet Allocation' does sound like a magical spell!
upvoted 0 times
...
Mollie
8 months ago
I agree, those algorithms are commonly used in machine learning.
upvoted 0 times
...
Gabriele
9 months ago
K-means and PCA are good choices for clustering and dimensionality reduction.
upvoted 0 times
...
...
Susy
10 months ago
Hmm, Semantic fegmentation? Is that a new algorithm or just a typo? I'd stick with the classics like K-means and Factorization Machines.
upvoted 0 times
Lillian
8 months ago
Definitely, K-means and Factorization Machines are tried and true algorithms for this task.
upvoted 0 times
...
Olen
9 months ago
Yeah, I agree. Those are solid choices for identifying customer groups.
upvoted 0 times
...
Quiana
9 months ago
I think Semantic fegmentation might be a typo, let's go with K-means and Factorization Machines.
upvoted 0 times
...
...
Tammy
10 months ago
Wow, 980 variables? That's a lot of data to work with! I'd go with K-means and PCA to start - get some nice clusters and reduce the dimensionality.
upvoted 0 times
Mariann
10 months ago
Agreed, K-means for clustering and PCA for dimensionality reduction make a good combination.
upvoted 0 times
...
Sylvie
10 months ago
K-means and PCA are great choices for handling such a large dataset.
upvoted 0 times
...
...
Nina
11 months ago
I'm not sure about PCA. Wouldn't Factorization machines (FM) be a better choice for this task?
upvoted 0 times
...
Justine
11 months ago
I agree with you, Maurine. K-means can help us cluster customers based on their similarities, and PCA can reduce the dimensionality of the dataset.
upvoted 0 times
...
Maurine
11 months ago
I think we should use K-means and Principal component analysis (PCA) for this task.
upvoted 0 times
...

Save Cancel