New Year Sale 2026! Hurry Up, Grab the Special Discount - Save 25% - Ends In 00:00:00 Coupon code: SAVE25
Welcome to Pass4Success

- Free Preparation Discussions

Amazon MLS-C01 Exam - Topic 2 Question 100 Discussion

Actual exam question for Amazon's MLS-C01 exam
Question #: 100
Topic #: 2
[All MLS-C01 Questions]

A data scientist uses Amazon SageMaker Data Wrangler to define and perform transformations and feature engineering on historical dat

a. The data scientist saves the transformations to SageMaker Feature Store.

The historical data is periodically uploaded to an Amazon S3 bucket. The data scientist needs to transform the new historic data and add it to the online feature store The data scientist needs to prepare the .....historic data for training and inference by using native integrations.

Which solution will meet these requirements with the LEAST development effort?

Show Suggested Answer Hide Answer
Suggested Answer: D

The best solution is to configure Amazon EventBridge to run a predefined SageMaker pipeline to perform the transformations when a new data is detected in the S3 bucket. This solution requires the least development effort because it leverages the native integration between EventBridge and SageMaker Pipelines, which allows you to trigger a pipeline execution based on an event rule. EventBridge can monitor the S3 bucket for new data uploads and invoke the pipeline that contains the same transformations and feature engineering steps that were defined in SageMaker Data Wrangler. The pipeline can then ingest the transformed data into the online feature store for training and inference.

The other solutions are less optimal because they require more development effort and additional services. Using AWS Lambda or AWS Step Functions would require writing custom code to invoke the SageMaker pipeline and handle any errors or retries. Using Apache Airflow would require setting up and maintaining an Airflow server and DAGs, as well as integrating with the SageMaker API.

References:

Amazon EventBridge and Amazon SageMaker Pipelines integration

Create a pipeline using a JSON specification

Ingest data into a feature group


Contribute your Thoughts:

0/2000 characters
Dottie
3 months ago
I disagree, using Apache Airflow feels unnecessary for this scenario.
upvoted 0 times
...
Timmy
3 months ago
A is pretty straightforward too, but I prefer D.
upvoted 0 times
...
Dannette
3 months ago
Wait, can EventBridge really trigger SageMaker pipelines? That's cool!
upvoted 0 times
...
Twanna
4 months ago
I think B might be overkill for this task.
upvoted 0 times
...
Nicolette
4 months ago
Option D seems like the easiest way to automate this!
upvoted 0 times
...
Alverta
4 months ago
I vaguely recall that Apache Airflow is great for orchestration, but it might require more setup than just using SageMaker directly.
upvoted 0 times
...
Corrie
4 months ago
I practiced a similar question, and I feel like AWS Step Functions could add complexity that isn't necessary here.
upvoted 0 times
...
Ciara
4 months ago
I think using Amazon EventBridge sounds familiar; it might be the easiest way to automate the pipeline when new data arrives.
upvoted 0 times
...
Leatha
5 months ago
I remember studying how AWS Lambda can trigger processes based on S3 events, but I'm not sure if it's the best option for this scenario.
upvoted 0 times
...
Georgiann
5 months ago
I'm leaning towards option B with Step Functions. The ability to orchestrate a series of transformations in a workflow seems well-suited for this use case. Plus, I'm more familiar with Step Functions, so I feel confident I could implement that solution efficiently.
upvoted 0 times
...
Kami
5 months ago
Hmm, I'm a bit unsure about this one. The question mentions "native integrations", so I'm wondering if one of the other options like Step Functions or Airflow might be a better fit to leverage those integrations. I'll need to think this through a bit more.
upvoted 0 times
...
Deja
5 months ago
This seems like a pretty straightforward question. I'd go with option A - using AWS Lambda to run a predefined SageMaker pipeline. It looks like the most direct way to automate the transformations on new data as it arrives.
upvoted 0 times
...
Esteban
5 months ago
Option D with EventBridge sounds interesting. I like the idea of having an event-driven approach to trigger the transformations automatically. That could help reduce the manual effort required. I'll make sure to read through the details of each choice carefully.
upvoted 0 times
...
Filiberto
5 months ago
Hmm, I'm a little unsure about this one. I know effective listening involves things like maintaining eye contact and nodding to show you're engaged, but I'm not totally sure which of these options is the best answer.
upvoted 0 times
...
Martha
1 year ago
I agree with Carey, Amazon EventBridge can trigger the transformations automatically with less effort.
upvoted 0 times
...
Lynette
1 year ago
Option D all the way! Easy peasy, just like my grandma's apple pie. Now, where's the feature store to put all those delicious data transformations?
upvoted 0 times
Jeannetta
1 year ago
Then you can prepare the historic data for training and inference using native integrations.
upvoted 0 times
...
Johnetta
1 year ago
Yeah, you can save the transformations to SageMaker Feature Store.
upvoted 0 times
...
Melissa
1 year ago
I think the feature store is in SageMaker, where you can save the transformations.
upvoted 0 times
...
Jill
1 year ago
Option D all the way! Easy peasy, just like my grandma's apple pie.
upvoted 0 times
...
...
Remona
1 year ago
But option A involves using AWS Lambda, which can automate the process.
upvoted 0 times
...
Carey
1 year ago
I disagree, I believe option D is more efficient.
upvoted 0 times
...
Remona
1 year ago
I think option A is the best choice.
upvoted 0 times
...
Audra
1 year ago
Personally, I'd go with Option D. Why complicate things with Lambda, Step Functions, or Airflow when EventBridge can do the job elegantly?
upvoted 0 times
Emeline
1 year ago
I agree, Option D with Amazon EventBridge seems like the most efficient choice here.
upvoted 0 times
...
Romana
1 year ago
Option D is definitely the way to go. It's the simplest solution for this scenario.
upvoted 0 times
...
...
Gennie
1 year ago
Option C with Apache Airflow seems a bit overkill for this use case. EventBridge is probably the simplest and most efficient solution.
upvoted 0 times
...
King
1 year ago
I'm partial to Option B with AWS Step Functions. It gives you a bit more control and visibility over the workflow compared to just using EventBridge.
upvoted 0 times
Bette
1 year ago
I agree, it provides more control and visibility over the workflow.
upvoted 0 times
...
Clorinda
1 year ago
Option B with AWS Step Functions sounds like a good choice.
upvoted 0 times
...
...
Coral
2 years ago
Option D looks like the most straightforward way to handle this. EventBridge can trigger the SageMaker pipeline automatically when new data arrives in the S3 bucket.
upvoted 0 times
Sabine
1 year ago
Yes, it definitely simplifies the workflow for the data scientist.
upvoted 0 times
...
Rikki
1 year ago
I agree, using Amazon EventBridge to automate the process seems efficient.
upvoted 0 times
...
Dorathy
1 year ago
Option D looks like the most straightforward way to handle this. EventBridge can trigger the SageMaker pipeline automatically when new data arrives in the S3 bucket.
upvoted 0 times
...
...

Save Cancel