Amazon MLA-C01 Übungsprüfungen
Zuletzt aktualisiert am 26.04.2025- Prüfungscode: MLA-C01
- Prüfungsname: AWS Certified Machine Learning Engineer - Associate
- Zertifizierungsanbieter: Amazon
- Zuletzt aktualisiert am: 26.04.2025
You are a data scientist at a retail company responsible for deploying a machine learning model that predicts customer purchase behavior. The model needs to serve real-time predictions with low latency to support the company’s recommendation engine on its e-commerce platform. The deployment solution must also be scalable to handle varying traffic loads during peak shopping periods, such as Black Friday and holiday sales. Additionally, you need to monitor the model’s performance and automatically roll out updates when a new version of the model is available.
Given these requirements, which AWS deployment service and configuration is the MOST SUITABLE for deploying the machine learning model?
- A . Deploy the model on Amazon EC2 instances with a load balancer to distribute traffic, manually scaling the instances based on expected traffic during peak periods
- B . Deploy the model on Amazon SageMaker with batch transform jobs, running the jobs periodically to generate predictions and storing the results in Amazon S3 for the recommendation engine
- C . Deploy the model using Amazon SageMaker real-time hosting services with an auto-scaling endpoint, enabling you to automatically adjust the number of instances based on traffic demand
- D . Use AWS Lambda to deploy the model as a serverless function, automatically scaling based on the number of requests, and store the model artifacts in Amazon S3
When is unexplainability not acceptable?
- A . When determining the result of a sports match
- B . When making product recommendations
- C . When explaining why a loan was declined
- D . When explaining why a transaction was deemed fraudulent
What is the bias versus variance trade-off in machine learning?
- A . The bias versus variance trade-off refers to the balance between underfitting and overfitting, where high bias leads to overfitting and high variance leads to underfitting
- B . The bias versus variance trade-off is a technique used to improve model performance by increasing both bias and variance simultaneously to achieve better generalization
- C . The bias versus variance trade-off refers to the challenge of balancing the error due to the model’s complexity (variance) and the error due to incorrect assumptions in the model (bias), where high bias can cause underfitting and high variance can cause overfitting
- D . The bias versus variance trade-off involves choosing between a model with high complexity that may capture more noise (high bias) and a simpler model that may generalize better but miss important patterns (high variance)
A company has recently migrated to AWS Cloud and it wants to optimize the hardware used for its AI workflows.
Which of the following would you suggest?
- A . Leverage either AWS Trainium or AWS Inferentia for the deep learning (DL) and generative AI inference applications
- B . Leverage AWS Trainium for high-performance, cost-effective Deep Learning training. Leverage AWS Inferentia for the deep learning (DL) and generative AI inference applications
- C . Leverage either AWS Trainium or AWS Inferentia for high-performance, cost-effective Deep Learning training
- D . Leverage AWS Inferentia for high-performance, cost-effective Deep Learning training. Leverage AWS Trainium for the deep learning (DL) and generative AI inference applications
What role will machine learning play in the future of technology?
- A . It will only be used in healthcare
- B . It will be boundless with applications in various fields
- C . It will become obsolete
- D . It will be limited to fraud detection
You are a data engineer responsible for monitoring the performance of a suite of machine learning models deployed across multiple environments at an e-commerce company. The models are used for various tasks, including recommendation engines, demand forecasting, and customer segmentation. To ensure that these models are performing optimally, you need to set up a centralized dashboard that allows stakeholders to monitor key performance metrics such as latency, accuracy, throughput, and resource utilization. The dashboard should be user-friendly, provide insights at a glance, and support both technical and non-technical users.
Which approach is MOST SUITABLE for setting up a dashboard to monitor the performance metrics of these ML models?
- A . Use Amazon QuickSight to create a visual dashboard that integrates data from Amazon CloudWatch Logs via Amazon S3, providing interactive charts and graphs that allow stakeholders to drill down into specific metrics as needed
- B . Create a custom web application that pulls data from CloudWatch Logs, generating real-time visualizations of performance metrics. Embed the application within the company’s internal portal for easy access
- C . Set up a CloudWatch Dashboard to aggregate and display performance metrics such as CPU utilization, memory usage, and latency from various services. Use Amazon SNS to send daily reports based on these metrics to stakeholders
- D . Use AWS Lambda to query CloudWatch Logs and send the results to an S3 bucket daily, where stakeholders can manually review the metrics and create their own visualizations using spreadsheet software
You are a Machine Learning Engineer at a healthcare company working on a binary classification model to predict whether a patient has a particular disease based on several medical features. The consequences of misclassifications are severe: false positives lead to unnecessary and expensive follow-up tests, while false negatives could result in a failure to provide critical treatment. You need to evaluate the model using appropriate metrics to balance the risks associated with these types of errors. Given the critical nature of the application, which combination of evaluation metrics should you prioritize to minimize both false positives and false negatives while ensuring that the model is reliable for deployment? (Select two)
- A . Prioritize accuracy, as it provides a general sense of how well the model is performing across all predictions
- B . Focus on precision to reduce the number of false positives, thus avoiding unnecessary follow-up tests for patients who do not have the disease
- C . Evaluate the model using the Area Under the ROC Curve (AUC) to understand its performance across different classification thresholds
- D . Use the F1 score to balance the trade-off between precision and recall, ensuring that both false positives and false negatives are considered
- E . Prioritize recall to reduce the number of false negatives, ensuring that as many patients with the disease as possible are correctly identified
Which solution will meet these requirements?
- A . Use the SageMaker Feature Store GetRecord API with the record identifier.
- B . Use the SageMaker Feature Store BatchGetRecord API with the record identifier. Filter to find the latest record.
- C . Create an Amazon Athena query to retrieve the data from the feature table. Use the write_time value to find the latest record.
- D . Create an Amazon Athena query to retrieve the data from the feature table.
How does the Amazon flywheel model ensure continuous improvement and growth?
- A . By using feedback loops that integrate customer experience, selection, and pricing.
- B . By limiting the scope of operational improvements to cost reduction only.
- C . By avoiding changes to business practices.
- D . By maintaining a static approach to business operations.
How does model training work in Deep Learning?
- A . Model training in deep learning involves only the use of support vector machines and decision trees to create predictive models
- B . Model training in deep learning involves using large datasets to adjust the weights and biases of a neural network through multiple iterations, using techniques such as gradient descent to minimize the error
- C . Model training in deep learning involves manually setting the weights and biases of a neural network based on predefined rules
- D . Model training in deep learning requires no data; the neural network automatically learns from
predefined algorithms without any input