Amazon MLS-C01 Dumps PDF

October, 2025 MLS-C01 Practice Questions

Make success possible with our Latest and Unique AWS Certified Specialty MLS-C01 Practice Exam!


Name: AWS Certified Machine Learning - Specialty
Exam Code: MLS-C01
Certification: AWS Certified Specialty
Vendor: Amazon
Total Questions: 330
Last Updated: September 30, 2025
950 Satisfied Customers

discount banner

$49 $68.6 Add To Cart

Last Week Results

37
Amazon MLS-C01 customers passed exam this week.
95%
Average Score in Real MLS-C01 Exam in Testing Centre.
93%
MLS-C01 Exam Questions came from DumpsGroup Material.



Unique Spoto Amazon MLS-C01 Practice Questions

Success is simply the result of the efforts you put into the preparation. We at Dumpsgroup wish to make that preparation a lot easier. The AWS Certified Machine Learning - Specialty MLS-C01 Practice Exam we offer is solely for best results. Our IT experts put in their blood and sweat into carefully selecting and compiling these unique Practice Questions. So, you can achieve your dreams of becoming a AWS Certified Specialty professional. Now is the time to press that big buy button and take the first step to a better and brighter future.

Passing the Amazon MLS-C01 exam is simpler if you have globally valid resources and Dumpsgroup provides you just that. Millions of customers come to us daily, leaving the platform happy and satisfied. Because we aim to provide you with AWS Certified Specialty Practice Questions aligned with the latest patterns of the AWS Certified Machine Learning - Specialty Exam. And not just that, our reliable customer services are 24 hours at your beck and call to support you in every way necessary. Order now to see the MLS-C01 Exam results you always desired.

2 Surefire Ways to Pass Amazon MLS-C01 Exam!

You must have heard about candidates failing in a large quantity and perhaps tried yourself and fail to pass AWS Certified Machine Learning - Specialty. It is best to try Dumpsgroup’s MLS-C01 Practice Questions this time around. Dumpsgroup not only provides an authentic, valid, and accurate resource for your preparation. They simplified the training by dividing it into two different formats for ease and comfort. Now you can get the Amazon MLS-C01 in both PDF and Online Test Engine formats. Choose whichever or both to start your AWS Certified Specialty certification exam preparation.

Furthermore, Dumpsgroup gives a hefty percentage off on these Spoto MLS-C01 Practice Exam by applying a simple discount code; when the actual price is already so cheap. The updates for the first three months, from the date of your purchase, are FREE. Our esteemed customers cannot stop singing praises of our Amazon MLS-C01 Practice Questions. That is because we offer only the questions with the highest possibility of appearing in the actual exam. Download the free demo and see for yourself.

The MLS-C01 Practice Exam for Achievers

We know you have been struggling to compete with your colleagues in your workplace. That is why we provide the MLS-C01 Practice Questions to let you gain the upper hand that you always wanted. These questions and answers are a thorough guide in a simple and exam-like format! That makes understanding and excelling in your field way lot easier. Our aim is not just to help to pass the AWS Certified Specialty Exam but to make a Amazon professional out of you. For that purpose, our MLS-C01 Practice Exams are the best choice.

Why You Choose Us:

  1. We can give you a million reasons to choose us for your AWS Certified Machine Learning - Specialty preparation. But we narrow down to the basics:
  2. Our Free MLS-C01 Practice Questions in the demo version are easily downloadable. A surefire way to ensure you are entrusting your training to a reliable resource is looking at it yourself.
  3. Online Test Engine & PDF: we give you two different methods to prepare your AWS Certified Specialty exam; MLS-C01 Practice Exam PDF and an online Test Engine version. Now you can advance your skills in the real-like exam practice environment. Choose the method that suits you best and prepare yourself for success.
  4. Safe & Secure Transaction: you can take it easy while buying your MLS-C01 Practice Questions. Dumpsgroup uses the latest and secure payment method to preserve our customer privacy and money. Our staff personnel have aligned capable security systems with high-end security technology. You know your details are safe with us because we never save them to avoid any inconvenience later.
  5. 24-hour customer support: you no longer have to worry about getting into trouble because our reliable customer care staff are active 24 hours to provide you support whenever you want.

MLS-C01 Practice Exam to Pass!

There are many resources available online for the preparation of the AWS Certified Machine Learning - Specialty Exam. But that does mean that all of them are reliable. When your future as a AWS Certified Specialty certified is at risk, you have got to think twice while choosing Amazon MLS-C01 Practice Questions. Dumpsgroup is not only a verified source of training material but has been in this business for years. In those years, we researched on MLS-C01 Practice Exam and came up with the best solution. So, you can trust that we know what we are doing. Moreover, we have joined hands with Amazon experts and professionals who are exceptional in their skills. And these experts approved our MLS-C01 Practice Questions for AWS Certified Machine Learning - Specialty preparation.

Sample Questions


MLS-C01 Sample Question 1


A large company has developed a B1 application that generates reports and dashboards
using data collected from various operational metrics The company wants to provide
executives with an enhanced experience so they can use natural language to get data from
the reports The company wants the executives to be able ask questions using written and
spoken interlaces
Which combination of services can be used to build this conversational interface? (Select
THREE)

A. Alexa for Business
B. Amazon Connect
C. Amazon Lex
D. Amazon Poly
E. Amazon Comprehend
F. Amazon Transcribe


ANSWER : C,E,F



MLS-C01 Sample Question 2


A data scientist has been running an Amazon SageMaker notebook instance for a few
weeks. During this time, a new version of Jupyter Notebook was released along with
additional software updates. The security team mandates that all running SageMaker
notebook instances use the latest security and software updates provided by SageMaker.
How can the data scientist meet these requirements?

A. Call the CreateNotebookInstanceLifecycleConfig API operation
B. Create a new SageMaker notebook instance and mount the Amazon Elastic Block Store(Amazon EBS) volume from the original instance
C. Stop and then restart the SageMaker notebook instance
D. Call the UpdateNotebookInstanceLifecycleConfig API operation


ANSWER : C



MLS-C01 Sample Question 3


A manufacturing company has structured and unstructured data stored in an Amazon S3
bucket. A Machine Learning Specialist wants to use SQL to run queries on this data.
Which solution requires the LEAST effort to be able to query this data?

A. Use AWS Data Pipeline to transform the data and Amazon RDS to run queries.
B. Use AWS Glue to catalogue the data and Amazon Athena to run queries.
C. Use AWS Batch to run ETL on the data and Amazon Aurora to run the queries.
D. Use AWS Lambda to transform the data and Amazon Kinesis Data Analytics to runqueries.


ANSWER : B



MLS-C01 Sample Question 4


A financial services company wants to automate its loan approval process by building a
machine learning (ML) model. Each loan data point contains credit history from a thirdparty
data source and demographic information about the customer. Each loan approval
prediction must come with a report that contains an explanation for why the customer was
approved for a loan or was denied for a loan. The company will use Amazon SageMaker to
build the model.
Which solution will meet these requirements with the LEAST development effort?

A. Use SageMaker Model Debugger to automatically debug the predictions, generate theexplanation, and attach the explanation report.
B. Use AWS Lambda to provide feature importance and partial dependence plots. Use theplots to generate and attach the explanation report.
C. Use SageMaker Clarify to generate the explanation report. Attach the report to thepredicted results.
D. Use custom Amazon Cloud Watch metrics to generate the explanation report. Attach thereport to the predicted results.


ANSWER : C



MLS-C01 Sample Question 5


A data scientist uses Amazon SageMaker Data Wrangler to define and perform
transformations and feature engineering on historical data. The data scientist saves the
transformations to SageMaker Feature Store.
The historical data is periodically uploaded to an Amazon S3 bucket. The data scientist
needs to transform the new historic data and add it to the online feature store The data
scientist needs to prepare the .....historic data for training and inference by using native
integrations.
Which solution will meet these requirements with the LEAST development effort?

A. Use AWS Lambda to run a predefined SageMaker pipeline to perform thetransformations on each new dataset that arrives in the S3 bucket.
B. Run an AWS Step Functions step and a predefined SageMaker pipeline to perform thetransformations on each new dalaset that arrives in the S3 bucket
C. Use Apache Airflow to orchestrate a set of predefined transformations on each newdataset that arrives in the S3 bucket.
D. Configure Amazon EventBridge to run a predefined SageMaker pipeline to perform thetransformations when a new data is detected in the S3 bucket.


ANSWER : D



MLS-C01 Sample Question 6


A data scientist at a financial services company used Amazon SageMaker to train and
deploy a model that predicts loan defaults. The model analyzes new loan applications and
predicts the risk of loan default. To train the model, the data scientist manually extracted
loan data from a database. The data scientist performed the model training and
deployment steps in a Jupyter notebook that is hosted on SageMaker Studio notebooks.
The model's prediction accuracy is decreasing over time. Which combination of slept in the
MOST operationally efficient way for the data scientist to maintain the model's accuracy?
(Select TWO.)

A. Use SageMaker Pipelines to create an automated workflow that extracts fresh data,trains the model, and deploys a new version of the model.
B. Configure SageMaker Model Monitor with an accuracy threshold to check for model drift.Initiate an Amazon CloudWatch alarm when the threshold is exceeded. Connect theworkflow in SageMaker Pipelines with the CloudWatch alarm to automatically initiateretraining.
C. Store the model predictions in Amazon S3 Create a daily SageMaker Processing jobthat reads the predictions from Amazon S3, checks for changes in model predictionaccuracy, and sends an email notification if a significant change is detected.
D. Rerun the steps in the Jupyter notebook that is hosted on SageMaker Studio notebooksto retrain the model and redeploy a new version of the model.
E. Export the training and deployment code from the SageMaker Studio notebooks into aPython script. Package the script into an Amazon Elastic Container Service (Amazon ECS)task that an AWS Lambda function can initiate.


ANSWER : A,B



MLS-C01 Sample Question 7


A manufacturing company has a production line with sensors that collect hundreds of
quality metrics. The company has stored sensor data and manual inspection results in a
data lake for several months. To automate quality control, the machine learning team must
build an automated mechanism that determines whether the produced goods are good
quality, replacement market quality, or scrap quality based on the manual inspection
results.
Which modeling approach will deliver the MOST accurate prediction of product quality?

A. Amazon SageMaker DeepAR forecasting algorithm
B. Amazon SageMaker XGBoost algorithm
C. Amazon SageMaker Latent Dirichlet Allocation (LDA) algorithm
D. A convolutional neural network (CNN) and ResNet


ANSWER : D



MLS-C01 Sample Question 8


A machine learning (ML) specialist is using the Amazon SageMaker DeepAR forecasting
algorithm to train a model on CPU-based Amazon EC2 On-Demand instances. The model
currently takes multiple hours to train. The ML specialist wants to decrease the training
time of the model.
Which approaches will meet this requirement7 (SELECT TWO )

A. Replace On-Demand Instances with Spot Instances
B. Configure model auto scaling dynamically to adjust the number of instancesautomatically.
C. Replace CPU-based EC2 instances with GPU-based EC2 instances.
D. Use multiple training instances.
E. Use a pre-trained version of the model. Run incremental training.


ANSWER : C,D



MLS-C01 Sample Question 9


A company is creating an application to identify, count, and classify animal images that are
uploaded to the company’s website. The company is using the Amazon SageMaker image
classification algorithm with an ImageNetV2 convolutional neural network (CNN). The
solution works well for most animal images but does not recognize many animal species
that are less common.
The company obtains 10,000 labeled images of less common animal species and stores
the images in Amazon S3. A machine learning (ML) engineer needs to incorporate the
images into the model by using Pipe mode in SageMaker.
Which combination of steps should the ML engineer take to train the model? (Choose two.)

A. Use a ResNet model. Initiate full training mode by initializing the network with randomweights.
B. Use an Inception model that is available with the SageMaker image classificationalgorithm.
C. Create a .lst file that contains a list of image files and corresponding class labels. Uploadthe .lst file to Amazon S3.
D. Initiate transfer learning. Train the model by using the images of less common species.
E. Use an augmented manifest file in JSON Lines format.


ANSWER : C,D



MLS-C01 Sample Question 10


A company operates large cranes at a busy port. The company plans to use machine
learning (ML) for predictive maintenance of the cranes to avoid unexpected breakdowns
and to improve productivity.
The company already uses sensor data from each crane to monitor the health of the
cranes in real time. The sensor data includes rotation speed, tension, energy consumption,
vibration, pressure, and …perature for each crane. The company contracts AWS ML
experts to implement an ML solution.
Which potential findings would indicate that an ML-based solution is suitable for this
scenario? (Select TWO.)

A. The historical sensor data does not include a significant number of data points andattributes for certain time periods.
B. The historical sensor data shows that simple rule-based thresholds can predict cranefailures.
C. The historical sensor data contains failure data for only one type of crane model that isin operation and lacks failure data of most other types of crane that are in operation.
D. The historical sensor data from the cranes are available with high granularity for the last3 years.
E. The historical sensor data contains most common types of crane failures that thecompany wants to predict.


ANSWER : D,E



MLS-C01 Sample Question 11


A company wants to forecast the daily price of newly launched products based on 3 years
of data for older product prices, sales, and rebates. The time-series data has irregular
timestamps and is missing some values.
Data scientist must build a dataset to replace the missing values. The data scientist needs
a solution that resamptes the data daily and exports the data for further modeling.
Which solution will meet these requirements with the LEAST implementation effort?

A. Use Amazon EMR Serveriess with PySpark.
B. Use AWS Glue DataBrew.
C. Use Amazon SageMaker Studio Data Wrangler.
D. Use Amazon SageMaker Studio Notebook with Pandas.


ANSWER : C



MLS-C01 Sample Question 12


A company wants to predict stock market price trends. The company stores stock market
data each business day in Amazon S3 in Apache Parquet format. The company stores 20
GB of data each day for each stock code.
A data engineer must use Apache Spark to perform batch preprocessing data
transformations quickly so the company can complete prediction jobs before the stock
market opens the next day. The company plans to track more stock market codes and
needs a way to scale the preprocessing data transformations.
Which AWS service or feature will meet these requirements with the LEAST development
effort over time?

A. AWS Glue jobs
B. Amazon EMR cluster
C. Amazon Athena
D. AWS Lambda


ANSWER : A



MLS-C01 Sample Question 13


An ecommerce company wants to use machine learning (ML) to monitor fraudulent
transactions on its website. The company is using Amazon SageMaker to research, train,
deploy, and monitor the ML models.
The historical transactions data is in a .csv file that is stored in Amazon S3 The data
contains features such as the user's IP address, navigation time, average time on each
page, and the number of clicks for ....session. There is no label in the data to indicate if a
transaction is anomalous.
Which models should the company use in combination to detect anomalous transactions?
(Select TWO.)

A. IP Insights
B. K-nearest neighbors (k-NN)
C. Linear learner with a logistic function
D. Random Cut Forest (RCF)
E. XGBoost


ANSWER : D,E



MLS-C01 Sample Question 14


A data scientist is working on a public sector project for an urban traffic system. While
studying the traffic patterns, it is clear to the data scientist that the traffic behavior at each
light is correlated, subject to a small stochastic error term. The data scientist must model
the traffic behavior to analyze the traffic patterns and reduce congestion.
How will the data scientist MOST effectively model the problem?

A. The data scientist should obtain a correlated equilibrium policy by formulating thisproblem as a multi-agent reinforcement learning problem.
B. The data scientist should obtain the optimal equilibrium policy by formulating thisproblem as a single-agent reinforcement learning problem.
C. Rather than finding an equilibrium policy, the data scientist should obtain accuratepredictors of traffic flow by using historical data through a supervised learning approach.
D. Rather than finding an equilibrium policy, the data scientist should obtain accuratepredictors of traffic flow by using unlabeled simulated data representing the new trafficpatterns in the city and applying an unsupervised learning approach.


ANSWER : A



MLS-C01 Sample Question 15


A company wants to create an artificial intelligence (Al) yoga instructor that can lead large
classes of students. The company needs to create a feature that can accurately count the
number of students who are in a class. The company also needs a feature that can
differentiate students who are performing a yoga stretch correctly from students who are
performing a stretch incorrectly.
...etermine whether students are performing a stretch correctly, the solution needs to
measure the location and angle of each student's arms and legs A data scientist must use
Amazon SageMaker to ...ss video footage of a yoga class by extracting image frames and
applying computer vision models.
Which combination of models will meet these requirements with the LEAST effort? (Select
TWO.)

A. Image Classification
B. Optical Character Recognition (OCR)
C. Object Detection
D. Pose estimation
E. Image Generative Adversarial Networks (GANs)


ANSWER : C,D



MLS-C01 Sample Question 16


A company wants to predict the classification of documents that are created from an
application. New documents are saved to an Amazon S3 bucket every 3 seconds. The
company has developed three versions of a machine learning (ML) model within Amazon
SageMaker to classify document text. The company wants to deploy these three versions to predict the classification of each document.
Which approach will meet these requirements with the LEAST operational overhead?

A. Configure an S3 event notification that invokes an AWS Lambda function when newdocuments are created. Configure the Lambda function to create three SageMaker batchtransform jobs, one batch transform job for each model for each document.
B. Deploy all the models to a single SageMaker endpoint. Treat each model as aproduction variant. Configure an S3 event notification that invokes an AWS Lambdafunction when new documents are created. Configure the Lambda function to call eachproduction variant and return the results of each model.
C. Deploy each model to its own SageMaker endpoint Configure an S3 event notificationthat invokes an AWS Lambda function when new documents are created. Configure theLambda function to call each endpoint and return the results of each model.
D. Deploy each model to its own SageMaker endpoint. Create three AWS Lambdafunctions. Configure each Lambda function to call a different endpoint and return theresults. Configure three S3 event notifications to invoke the Lambda functions when newdocuments are created.


ANSWER : B



MLS-C01 Sample Question 17


A company is using Amazon Polly to translate plaintext documents to speech for
automated company announcements However company acronyms are being
mispronounced in the current documents How should a Machine Learning Specialist
address this issue for future documents?

A. Convert current documents to SSML with pronunciation tags
B. Create an appropriate pronunciation lexicon.
C. Output speech marks to guide in pronunciation
D. Use Amazon Lex to preprocess the text files for pronunciation


ANSWER : B



MLS-C01 Sample Question 18


An online delivery company wants to choose the fastest courier for each delivery at the
moment an order is placed. The company wants to implement this feature for existing users
and new users of its application. Data scientists have trained separate models with
XGBoost for this purpose, and the models are stored in Amazon S3. There is one model fof
each city where the company operates.
The engineers are hosting these models in Amazon EC2 for responding to the web client
requests, with one instance for each model, but the instances have only a 5% utilization in
CPU and memory, ....operation engineers want to avoid managing unnecessary resources.
Which solution will enable the company to achieve its goal with the LEAST operational
overhead?

A. Create an Amazon SageMaker notebook instance for pulling all the models fromAmazon S3 using the boto3 library. Remove the existing instances and use the notebook toperform a SageMaker batch transform for performing inferences offline for all the possibleusers in all the cities. Store the results in different files in Amazon S3. Point the web clientto the files.
B. Prepare an Amazon SageMaker Docker container based on the open-source multimodelserver. Remove the existing instances and create a multi-model endpoint inSageMaker instead, pointing to the S3 bucket containing all the models Invoke theendpoint from the web client at runtime, specifying the TargetModel parameter according tothe city of each request.
C. Keep only a single EC2 instance for hosting all the models. Install a model server in theinstance and load each model by pulling it from Amazon S3. Integrate the instance with theweb client using Amazon API Gateway for responding to the requests in real time,specifying the target resource according to the city of each request.
D. Prepare a Docker container based on the prebuilt images in Amazon SageMaker.Replace the existing instances with separate SageMaker endpoints. one for each citywhere the company operates. Invoke the endpoints from the web client, specifying the URL and EndpomtName parameter according to the city of each request.


ANSWER : B



MLS-C01 Sample Question 19


A data engineer is preparing a dataset that a retail company will use to predict the number
of visitors to stores. The data engineer created an Amazon S3 bucket. The engineer
subscribed the S3 bucket to an AWS Data Exchange data product for general economic
indicators. The data engineer wants to join the economic indicator data to an existing table
in Amazon Athena to merge with the business data. All these transformations must finish
running in 30-60 minutes.
Which solution will meet these requirements MOST cost-effectively?

A. Configure the AWS Data Exchange product as a producer for an Amazon Kinesis datastream. Use an Amazon Kinesis Data Firehose delivery stream to transfer the data toAmazon S3 Run an AWS Glue job that will merge the existing business data with theAthena table. Write the result set back to Amazon S3.
B. Use an S3 event on the AWS Data Exchange S3 bucket to invoke an AWS Lambdafunction. Program the Lambda function to use Amazon SageMaker Data Wrangler tomerge the existing business data with the Athena table. Write the result set back toAmazon S3.
C. Use an S3 event on the AWS Data Exchange S3 bucket to invoke an AWS LambdaFunction Program the Lambda function to run an AWS Glue job that will merge the existingbusiness data with the Athena table Write the results back to Amazon S3.
D. Provision an Amazon Redshift cluster. Subscribe to the AWS Data Exchange productand use the product to create an Amazon Redshift Table Merge the data in AmazonRedshift. Write the results back to Amazon S3.


ANSWER : B



MLS-C01 Sample Question 20


A machine learning (ML) specialist is using Amazon SageMaker hyperparameter
optimization (HPO) to improve a model’s accuracy. The learning rate parameter is specified
in the following HPO configuration:


During the results analysis, the ML specialist determines that most of the training jobs had
a learning rate between 0.01 and 0.1. The best result had a learning rate of less than 0.01.
Training jobs need to run regularly over a changing dataset. The ML specialist needs to
find a tuning mechanism that uses different learning rates more evenly from the provided
range between MinValue and MaxValue.
Which solution provides the MOST accurate result?

A.Modify the HPO configuration as follows: Select the most accurate hyperparameter configuration form this HPO job.
B.Run three different HPO jobs that use different learning rates form the following intervalsfor MinValue and MaxValue while using the same number of training jobs for each HPOjob:[0.01, 0.1][0.001, 0.01][0.0001, 0.001]Select the most accurate hyperparameter configuration form these three HPO jobs.
C.Modify the HPO configuration as follows: Select the most accurate hyperparameter configuration form this training job.
D.Run three different HPO jobs that use different learning rates form the following intervalsfor MinValue and MaxValue. Divide the number of training jobs for each HPO job by three:[0.01, 0.1][0.001, 0.01][0.0001, 0.001]Select the most accurate hyperparameter configuration form these three HPO jobs.


ANSWER : C



MLS-C01 Sample Question 21


A retail company wants to build a recommendation system for the company's website. The
system needs to provide recommendations for existing users and needs to base those
recommendations on each user's past browsing history. The system also must filter out any
items that the user previously purchased.
Which solution will meet these requirements with the LEAST development effort?

A. Train a model by using a user-based collaborative filtering algorithm on AmazonSageMaker. Host the model on a SageMaker real-time endpoint. Configure an Amazon APIGateway API and an AWS Lambda function to handle real-time inference requests that theweb application sends. Exclude the items that the user previously purchased from theresults before sending the results back to the web application.
B. Use an Amazon Personalize PERSONALIZED_RANKING recipe to train a model.Create a real-time filter to exclude items that the user previously purchased. Create anddeploy a campaign on Amazon Personalize. Use the GetPersonalizedRanking APIoperation to get the real-time recommendations.
C. Use an Amazon Personalize USER_ PERSONAL IZATION recipe to train a modelCreate a real-time filter to exclude items that the user previously purchased. Create anddeploy a campaign on Amazon Personalize. Use the GetRecommendations API operationto get the real-time recommendations.
D. Train a neural collaborative filtering model on Amazon SageMaker by using GPU instances. Host the model on a SageMaker real-time endpoint. Configure an Amazon APIGateway API and an AWS Lambda function to handle real-time inference requests that theweb application sends. Exclude the items that the user previously purchased from theresults before sending the results back to the web application.


ANSWER : C



MLS-C01 Sample Question 22


A company deployed a machine learning (ML) model on the company website to predict
real estate prices. Several months after deployment, an ML engineer notices that the
accuracy of the model has gradually decreased.
The ML engineer needs to improve the accuracy of the model. The engineer also needs to
receive notifications for any future performance issues.
Which solution will meet these requirements?

A. Perform incremental training to update the model. Activate Amazon SageMaker Model Monitor to detect model performance issues and to send notifications.
B. Use Amazon SageMaker Model Governance. Configure Model Governance toautomatically adjust model hyper para meters. Create a performance threshold alarm inAmazon CloudWatch to send notifications.
C. Use Amazon SageMaker Debugger with appropriate thresholds. Configure Debugger tosend Amazon CloudWatch alarms to alert the team Retrain the model by using only datafrom the previous several months.
D. Use only data from the previous several months to perform incremental training toupdate the model. Use Amazon SageMaker Model Monitor to detect model performanceissues and to send notifications.


ANSWER : A



MLS-C01 Sample Question 23


A company wants to detect credit card fraud. The company has observed that an average
of 2% of credit card transactions are fraudulent. A data scientist trains a classifier on a
year's worth of credit card transaction data. The classifier needs to identify the fraudulent
transactions. The company wants to accurately capture as many fraudulent transactions as
possible.
Which metrics should the data scientist use to optimize the classifier? (Select TWO.)

A. Specificity
B. False positive rate
C. Accuracy
D. Fl score
E. True positive rate


ANSWER : D,E



MLS-C01 Sample Question 24


An insurance company developed a new experimental machine learning (ML) model to
replace an existing model that is in production. The company must validate the quality of
predictions from the new experimental model in a production environment before the
company uses the new experimental model to serve general user requests.
Which one model can serve user requests at a time. The company must measure the
performance of the new experimental model without affecting the current live traffic
Which solution will meet these requirements?

A. A/B testing
B. Canary release
C. Shadow deployment
D. Blue/green deployment


ANSWER : C



MLS-C01 Sample Question 25


A university wants to develop a targeted recruitment strategy to increase new student
enrollment. A data scientist gathers information about the academic performance history of
students. The data scientist wants to use the data to build student profiles. The university
will use the profiles to direct resources to recruit students who are likely to enroll in the
university.
Which combination of steps should the data scientist take to predict whether a particular
student applicant is likely to enroll in the university? (Select TWO)

A. Use Amazon SageMaker Ground Truth to sort the data into two groups named"enrolled" or "not enrolled."
B. Use a forecasting algorithm to run predictions.
C. Use a regression algorithm to run predictions.
D. Use a classification algorithm to run predictions
E. Use the built-in Amazon SageMaker k-means algorithm to cluster the data into twogroups named "enrolled" or "not enrolled."


ANSWER : A,D



MLS-C01 Sample Question 26


A car company is developing a machine learning solution to detect whether a car is present
in an image. The image dataset consists of one million images. Each image in the dataset
is 200 pixels in height by 200 pixels in width. Each image is labeled as either having a car
or not having a car.
Which architecture is MOST likely to produce a model that detects whether a car is present
in an image with the highest accuracy?

A. Use a deep convolutional neural network (CNN) classifier with the images as input.Include a linear output layer that outputs the probability that an image contains a car.
B. Use a deep convolutional neural network (CNN) classifier with the images as input.Include a softmax output layer that outputs the probability that an image contains a car.
C. Use a deep multilayer perceptron (MLP) classifier with the images as input. Include alinear output layer that outputs the probability that an image contains a car.
D. Use a deep multilayer perceptron (MLP) classifier with the images as input. Include asoftmax output layer that outputs the probability that an image contains a car.


ANSWER : A



MLS-C01 Sample Question 27


A data scientist is building a linear regression model. The scientist inspects the dataset and
notices that the mode of the distribution is lower than the median, and the median is lower
than the mean.
Which data transformation will give the data scientist the ability to apply a linear regression
model?

A. Exponential transformation
B. Logarithmic transformation
C. Polynomial transformation
D. Sinusoidal transformation


ANSWER : B



MLS-C01 Sample Question 28


A network security vendor needs to ingest telemetry data from thousands of endpoints that
run all over the world. The data is transmitted every 30 seconds in the form of records that
contain 50 fields. Each record is up to 1 KB in size. The security vendor uses Amazon
Kinesis Data Streams to ingest the data. The vendor requires hourly summaries of the
records that Kinesis Data Streams ingests. The vendor will use Amazon Athena to query
the records and to generate the summaries. The Athena queries will target 7 to 12 of the
available data fields.
Which solution will meet these requirements with the LEAST amount of customization to
transform and store the ingested data?

A. Use AWS Lambda to read and aggregate the data hourly. Transform the data and storeit in Amazon S3 by using Amazon Kinesis Data Firehose.
B. Use Amazon Kinesis Data Firehose to read and aggregate the data hourly. Transformthe data and store it in Amazon S3 by using a short-lived Amazon EMR cluster.
C. Use Amazon Kinesis Data Analytics to read and aggregate the data hourly. Transformthe data and store it in Amazon S3 by using Amazon Kinesis Data Firehose.
D. Use Amazon Kinesis Data Firehose to read and aggregate the data hourly. Transform the data and store it in Amazon S3 by using AWS Lambda.


ANSWER : C



MLS-C01 Sample Question 29


A machine learning (ML) engineer has created a feature repository in Amazon SageMaker
Feature Store for the company. The company has AWS accounts for development,
integration, and production. The company hosts a feature store in the development
account. The company uses Amazon S3 buckets to store feature values offline. The
company wants to share features and to allow the integration account and the production
account to reuse the features that are in the feature repository. Which combination of steps will meet these requirements? (Select TWO.)

A. Create an IAM role in the development account that the integration account andproduction account can assume. Attach IAM policies to the role that allow access to thefeature repository and the S3 buckets.
B. Share the feature repository that is associated the S3 buckets from the developmentaccount to the integration account and the production account by using AWS ResourceAccess Manager (AWS RAM).
C. Use AWS Security Token Service (AWS STS) from the integration account and theproduction account to retrieve credentials for the development account.
D. Set up S3 replication between the development S3 buckets and the integration andproduction S3 buckets.
E. Create an AWS PrivateLink endpoint in the development account for SageMaker.


ANSWER : A,B



MLS-C01 Sample Question 30


An Amazon SageMaker notebook instance is launched into Amazon VPC The SageMaker
notebook references data contained in an Amazon S3 bucket in another account The
bucket is encrypted using SSE-KMS The instance returns an access denied error when
trying to access data in Amazon S3.
Which of the following are required to access the bucket and avoid the access denied
error? (Select THREE)

A. An AWS KMS key policy that allows access to the customer master key (CMK)
B. A SageMaker notebook security group that allows access to Amazon S3
C. An 1AM role that allows access to the specific S3 bucket
D. A permissive S3 bucket policy
E. An S3 bucket owner that matches the notebook owner
F. A SegaMaker notebook subnet ACL that allow traffic to Amazon S3.


ANSWER : A,B,C



MLS-C01 Sample Question 31


An engraving company wants to automate its quality control process for plaques. The
company performs the process before mailing each customized plaque to a customer. The
company has created an Amazon S3 bucket that contains images of defects that should
cause a plaque to be rejected. Low-confidence predictions must be sent to an internal team
of reviewers who are using Amazon Augmented Al (Amazon A2I).
Which solution will meet these requirements?

A. Use Amazon Textract for automatic processing. Use Amazon A2I with AmazonMechanical Turk for manual review.
B. Use Amazon Rekognition for automatic processing. Use Amazon A2I with a privateworkforce option for manual review.
C. Use Amazon Transcribe for automatic processing. Use Amazon A2I with a privateworkforce option for manual review.
D. Use AWS Panorama for automatic processing Use Amazon A2I with AmazonMechanical Turk for manual review


ANSWER : B



MLS-C01 Sample Question 32


A company builds computer-vision models that use deep learning for the autonomous
vehicle industry. A machine learning (ML) specialist uses an Amazon EC2 instance that
has a CPU: GPU ratio of 12:1 to train the models.
The ML specialist examines the instance metric logs and notices that the GPU is idle half of
the time The ML specialist must reduce training costs without increasing the duration of the
training jobs.
Which solution will meet these requirements?

A. Switch to an instance type that has only CPUs.
B. Use a heterogeneous cluster that has two different instances groups.
C. Use memory-optimized EC2 Spot Instances for the training jobs.
D. Switch to an instance type that has a CPU GPU ratio of 6:1.


ANSWER : D



MLS-C01 Sample Question 33


A data scientist is training a large PyTorch model by using Amazon SageMaker. It takes 10
hours on average to train the model on GPU instances. The data scientist suspects that
training is not converging and that
resource utilization is not optimal.
What should the data scientist do to identify and address training issues with the LEAST
development effort?

A. Use CPU utilization metrics that are captured in Amazon CloudWatch. Configure aCloudWatch alarm to stop the training job early if low CPU utilization occurs.
B. Use high-resolution custom metrics that are captured in Amazon CloudWatch. Configurean AWS Lambda function to analyze the metrics and to stop the training job early if issuesare detected.
C. Use the SageMaker Debugger vanishing_gradient and LowGPUUtilization built-in rulesto detect issues and to launch the StopTrainingJob action if issues are detected.
D. Use the SageMaker Debugger confusion and feature_importance_overweight built-inrules to detect issues and to launch the StopTrainingJob action if issues are detected.


ANSWER : C



MLS-C01 Sample Question 34


A data scientist is building a forecasting model for a retail company by using the most
recent 5 years of sales records that are stored in a data warehouse. The dataset contains
sales records for each of the company's stores across five commercial regions The data
scientist creates a working dataset with StorelD. Region. Date, and Sales Amount as
columns. The data scientist wants to analyze yearly average sales for each region. The
scientist also wants to compare how each region performed compared to average sales
across all commercial regions.
Which visualization will help the data scientist better understand the data trend?

A. Create an aggregated dataset by using the Pandas GroupBy function to get averagesales for each year for each store. Create a bar plot, faceted by year, of average sales foreach store. Add an extra bar in each facet to represent average sales.
B. Create an aggregated dataset by using the Pandas GroupBy function to get averagesales for each year for each store. Create a bar plot, colored by region and faceted by year,of average sales for each store. Add a horizontal line in each facet to represent averagesales.
C. Create an aggregated dataset by using the Pandas GroupBy function to get averagesales for each year for each region Create a bar plot of average sales for each region. Addan extra bar in each facet to represent average sales.
D. Create an aggregated dataset by using the Pandas GroupBy function to get average sales for each year for each region Create a bar plot, faceted by year, of average sales foreach region Add a horizontal line in each facet to represent average sales.


ANSWER : D



MLS-C01 Sample Question 35


A media company is building a computer vision model to analyze images that are on social
media. The model consists of CNNs that the company trained by using images that the
company stores in Amazon S3. The company used an Amazon SageMaker training job in
File mode with a single Amazon EC2 On-Demand Instance.
Every day, the company updates the model by using about 10,000 images that the
company has collected in the last 24 hours. The company configures training with only one
epoch. The company wants to speed up training and lower costs without the need to make
any code changes.
Which solution will meet these requirements?

A. Instead of File mode, configure the SageMaker training job to use Pipe mode. Ingest thedata from a pipe.
B. Instead Of File mode, configure the SageMaker training job to use FastFile mode withno Other changes.
C. Instead Of On-Demand Instances, configure the SageMaker training job to use SpotInstances. Make no Other changes.
D. Instead Of On-Demand Instances, configure the SageMaker training job to use SpotInstances. Implement model checkpoints.


ANSWER : C



MLS-C01 Sample Question 36


An automotive company uses computer vision in its autonomous cars. The company
trained its object detection models successfully by using transfer learning from a
convolutional neural network (CNN). The company trained the models by using PyTorch through the Amazon SageMaker SDK.
The vehicles have limited hardware and compute power. The company wants to optimize
the model to reduce memory, battery, and hardware consumption without a significant
sacrifice in accuracy.
Which solution will improve the computational efficiency of the models?

A. Use Amazon CloudWatch metrics to gain visibility into the SageMaker training weights,gradients, biases, and activation outputs. Compute the filter ranks based on the traininginformation. Apply pruning to remove the low-ranking filters. Set new weights based on thepruned set of filters. Run a new training job with the pruned model.
B. Use Amazon SageMaker Ground Truth to build and run data labeling workflows. Collecta larger labeled dataset with the labelling workflows. Run a new training job that uses thenew labeled data with previous training data.
C. Use Amazon SageMaker Debugger to gain visibility into the training weights, gradients,biases, and activation outputs. Compute the filter ranks based on the training information.Apply pruning to remove the low-ranking filters. Set the new weights based on the prunedset of filters. Run a new training job with the pruned model.
D. Use Amazon SageMaker Model Monitor to gain visibility into the ModelLatency metricand OverheadLatency metric of the model after the company deploys the model. Increasethe model learning rate. Run a new training job.


ANSWER : C



MLS-C01 Sample Question 37


A retail company stores 100 GB of daily transactional data in Amazon S3 at periodic
intervals. The company wants to identify the schema of the transactional data. The
company also wants to perform transformations on the transactional data that is in Amazon
S3.
The company wants to use a machine learning (ML) approach to detect fraud in the
transformed data.
Which combination of solutions will meet these requirements with the LEAST operational
overhead? {Select THREE.)

A. Use Amazon Athena to scan the data and identify the schema.
B. Use AWS Glue crawlers to scan the data and identify the schema.
C. Use Amazon Redshift to store procedures to perform data transformations
D. Use AWS Glue workflows and AWS Glue jobs to perform data transformations.
E. Use Amazon Redshift ML to train a model to detect fraud.
F. Use Amazon Fraud Detector to train a model to detect fraud.


ANSWER : B,D,F



MLS-C01 Sample Question 38


A media company wants to create a solution that identifies celebrities in pictures that users
upload. The company also wants to identify the IP address and the timestamp details from
the users so the company can prevent users from uploading pictures from unauthorized
locations.
Which solution will meet these requirements with LEAST development effort?

A. Use AWS Panorama to identify celebrities in the pictures. Use AWS CloudTrail tocapture IP address and timestamp details.
B. Use AWS Panorama to identify celebrities in the pictures. Make calls to the AWSPanorama Device SDK to capture IP address and timestamp details.
C. Use Amazon Rekognition to identify celebrities in the pictures. Use AWS CloudTrail tocapture IP address and timestamp details.
D. Use Amazon Rekognition to identify celebrities in the pictures. Use the text detectionfeature to capture IP address and timestamp details.


ANSWER : C



MLS-C01 Sample Question 39


A pharmaceutical company performs periodic audits of clinical trial sites to quickly resolve
critical findings. The company stores audit documents in text format. Auditors have
requested help from a data science team to quickly analyze the documents. The auditors
need to discover the 10 main topics within the documents to prioritize and distribute the
review work among the auditing team members. Documents that describe adverse events
must receive the highest priority. A data scientist will use statistical modeling to discover abstract topics and to provide a list
of the top words for each category to help the auditors assess the relevance of the topic.
Which algorithms are best suited to this scenario? (Choose two.)

A. Latent Dirichlet allocation (LDA)
B. Random Forest classifier
C. Neural topic modeling (NTM)
D. Linear support vector machine
E. Linear regression


ANSWER : A,C



MLS-C01 Sample Question 40


A credit card company wants to identify fraudulent transactions in real time. A data scientist
builds a machine learning model for this purpose. The transactional data is captured and
stored in Amazon S3. The historic data is already labeled with two classes: fraud (positive)
and fair transactions (negative). The data scientist removes all the missing data and builds
a classifier by using the XGBoost algorithm in Amazon SageMaker. The model produces
the following results:
• True positive rate (TPR): 0.700
• False negative rate (FNR): 0.300
• True negative rate (TNR): 0.977
• False positive rate (FPR): 0.023
• Overall accuracy: 0.949
Which solution should the data scientist use to improve the performance of the model?

A. Apply the Synthetic Minority Oversampling Technique (SMOTE) on the minority class inthe training dataset. Retrain the model with the updated training data.
B. Apply the Synthetic Minority Oversampling Technique (SMOTE) on the majority class in the training dataset. Retrain the model with the updated training data.
C. Undersample the minority class.
D. Oversample the majority class.


ANSWER : A



MLS-C01 Sample Question 41


A Machine Learning Specialist is designing a scalable data storage solution for Amazon
SageMaker. There is an existing TensorFlow-based model implemented as a train.py script
that relies on static training data that is currently stored as TFRecords.
Which method of providing training data to Amazon SageMaker would meet the business
requirements with the LEAST development overhead?

A. Use Amazon SageMaker script mode and use train.py unchanged. Point the AmazonSageMaker training invocation to the local path of the data without reformatting the trainingdata.
B. Use Amazon SageMaker script mode and use train.py unchanged. Put the TFRecorddata into an Amazon S3 bucket. Point the Amazon SageMaker training invocation to the S3bucket without reformatting the training data.
C. Rewrite the train.py script to add a section that converts TFRecords to protobuf andingests the protobuf data instead of TFRecords.
D. Prepare the data in the format accepted by Amazon SageMaker. Use AWS Glue orAWS Lambda to reformat and store the data in an Amazon S3 bucket.


ANSWER : B



MLS-C01 Sample Question 42


A data scientist stores financial datasets in Amazon S3. The data scientist uses Amazon
Athena to query the datasets by using SQL.
The data scientist uses Amazon SageMaker to deploy a machine learning (ML) model. The
data scientist wants to obtain inferences from the model at the SageMaker endpoint
However, when the data …. ntist attempts to invoke the SageMaker endpoint, the data
scientist receives SOL statement failures The data scientist's 1AM user is currently unable
to invoke the SageMaker endpoint
Which combination of actions will give the data scientist's 1AM user the ability to invoke the SageMaker endpoint? (Select THREE.)

A. Attach the AmazonAthenaFullAccess AWS managed policy to the user identity.
B. Include a policy statement for the data scientist's 1AM user that allows the 1AM user toperform the sagemaker: lnvokeEndpoint action,
C. Include an inline policy for the data scientist’s 1AM user that allows SageMaker to readS3 objects
D. Include a policy statement for the data scientist's 1AM user that allows the 1AM user toperform the sagemakerGetRecord action.
E. Include the SQL statement "USING EXTERNAL FUNCTION ml_function_name" in theAthena SQL query.
F. Perform a user remapping in SageMaker to map the 1AM user to another 1AM user thatis on the hosted endpoint.


ANSWER : B,C,E