Amazon MLA-C01 Übungsprüfungen
Zuletzt aktualisiert am 27.04.2025- Prüfungscode: MLA-C01
- Prüfungsname: AWS Certified Machine Learning Engineer - Associate
- Zertifizierungsanbieter: Amazon
- Zuletzt aktualisiert am: 27.04.2025
You are a data scientist working on a regression model to predict housing prices in a large metropolitan area. The dataset contains many features, including location, square footage, number of bedrooms, and amenities. After initial testing, you notice that some features have very high variance, leading to overfitting. To address this, you are considering applying regularization to your model. You need to choose between L1 (Lasso) and L2 (Ridge) regularization.
Given the goal of reducing overfitting while also simplifying the model by eliminating less important features, which regularization method should you choose and why?
- A . L2 regularization, because it evenly reduces all feature coefficients, leading to a more stable model
- B . L2 regularization, because it eliminates less important features by setting their coefficients to zero, simplifying the model
- C . L1 regularization, because it can shrink some feature coefficients to zero, effectively performing feature selection
- D . L1 regularization, because it penalizes large coefficients more heavily, making the model less
sensitive to high-variance features
You are responsible for deploying a machine learning model on AWS SageMaker for a real-time prediction application. The application requires low latency and high throughput. During deployment, you notice that the model’s response time is slower than expected, and the throughput is not meeting the required levels. You have already optimized the model itself, so the next step is to optimize the deployment environment. You are currently using a single instance of the ml.m5.large instance type with the default endpoint configuration.
Which of the following changes is MOST LIKELY to improve the model’s response time and throughput?
- A . Change the instance type to ml.p2.xlarge and add multi-model support
- B . Enable Auto Scaling with a target metric for the instance utilization
- C . Switch to an ml.m5.2xlarge instance type and use multi-AZ deployment
- D . Increase the instance count to two and enable asynchronous inference
Your data science team is working on developing a machine learning model to predict customer churn. The dataset that you are using contains hundreds of features, but you suspect that not all of these features are equally important for the model’s accuracy. To improve the model’s performance and reduce its complexity, the team wants to focus on selecting only the most relevant features that contribute significantly to minimizing the model’s error rate.
Which feature engineering process should your team apply to select a subset of features that are the most relevant towards minimizing the error rate of the trained model?
- A . Feature extraction
- B . Feature creation
- C . Feature transformation
- D . Feature selection
You are a data scientist working on a predictive maintenance model for an industrial manufacturing company. The model is designed to predict equipment failures based on sensor data collected over time. During the development process, you notice that the model performs exceptionally well on the training data but struggles to generalize to new, unseen data. Additionally, there are some indications that the model might not be fully capturing the complexity of the problem. To ensure the model performs well in production, you need to identify whether it is overfitting, underfitting, or both.
Which of the following strategies is the MOST EFFECTIVE for identifying overfitting and underfitting in your model?
- A . Perform cross-validation with different subsets of the data; if the model’s performance varies significantly across folds, the model is underfitting
- B . Compare the training and validation loss curves over time; if the validation loss is much higher than the training loss, the model is likely overfitting
- C . Reduce the number of features in the model; if performance improves, the model was previously overfitting
- D . Analyze the model’s performance on a separate test set; if the model performs well on both the training and test sets, it is neither overfitting nor underfitting
You are a machine learning engineer tasked with building a deep learning model to classify images for an autonomous vehicle project. The dataset is massive, consisting of millions of labeled images. Initial training runs on a single GPU instance in Amazon SageMaker are taking too long, and the training costs are rising. You need to reduce the model training time without compromising performance significantly.
Which of the following approaches is the MOST LIKELY to effectively reduce the training time while maintaining model performance?
- A . Implement distributed training using multiple GPU instances to parallelize the training process, reducing the overall time
- B . Reduce the size of the training dataset to speed up training, even if it means using fewer examples per class
- C . Switch to a smaller instance type to reduce computational costs, accepting a longer training time as a trade-off
- D . Enable early stopping to halt training when the model’s performance on the validation set stops improving, thereby avoiding overfitting
In what scenario would traditional programming techniques be preferable over ML?
- A . When complex logic and scalability are required.
- B . When the problem can be solved with simple rules.
- C . When personalized recommendations are needed.
- D . When quick adaptation to new data is necessary.
You are tasked with building a predictive model for customer lifetime value (CLV) using Amazon SageMaker. Given the complexity of the model, it’s crucial to optimize hyperparameters to achieve the best possible performance. You decide to use SageMaker’s automatic model tuning (hyperparameter optimization) with Random Search strategy to fine-tune the model. You have a large dataset, and the tuning job involves several hyperparameters, including the learning rate, batch size, and dropout rate. During the tuning process, you observe that some of the trials are not converging effectively, and the results are not as expected. You suspect that the hyperparameter ranges or the strategy you are using may need adjustment.
Which of the following approaches is MOST LIKELY to improve the effectiveness of the hyperparameter tuning process?
- A . Decrease the number of total trials but increase the number of parallel jobs to speed up the tuning process
- B . Switch from the Random Search strategy to the Bayesian Optimization strategy and narrow the range of critical hyperparameters
- C . Use the Grid Search strategy with a wide range for all hyperparameters and increase the number of total trials
- D . Increase the number of hyperparameters being tuned and widen the range for all hyperparameters
You are an ML engineer at a startup that is developing a recommendation engine for an e-commerce platform. The workload involves training models on large datasets and deploying them to serve real-time recommendations to customers. The training jobs are sporadic but require significant computational power, while the inference workloads must handle varying traffic throughout the day. The company is cost-conscious and aims to balance cost efficiency with the need for scalability and performance. Given these requirements, which approach to resource allocation is the MOST SUITABLE for training and inference, and why?
- A . Use on-demand instances for both training and inference to ensure that the company only pays for the compute resources it uses when it needs them, avoiding any upfront commitments
- B . Use on-demand instances for training, allowing the flexibility to scale resources as needed, and use provisioned resources with auto-scaling for inference to handle varying traffic while controlling costs
- C . Use provisioned resources with spot instances for both training and inference to take advantage of the lowest possible costs, accepting the potential for interruptions during workload execution
- D . Use provisioned resources with reserved instances for both training and inference to lock in lower
costs and guarantee resource availability, ensuring predictability in budgeting
Which of the following summarizes the differences between a token and an embedding in the context of generative AI?
- A . An embedding is a sequence of characters that a model can interpret or predict as a single unit of meaning, whereas, a token is a vector of numerical values that represents condensed information obtained by transforming input into that vector
- B . Both token and embedding refer to a sequence of characters that a model can interpret or predict as a single unit of meaning
- C . A token is a sequence of characters that a model can interpret or predict as a single unit of meaning, whereas, an embedding is a vector of numerical values that represents condensed information obtained by transforming input into that vector
- D . Both token and embedding refer to a vector of numerical values that represents condensed
information obtained by transforming input into that vector
A machine learning specialist is developing a proof of concept for government users whose primary concern is security. The specialist is using Amazon SageMaker to train a convolutional neural network (CNN) model for a photo classifier application. The specialist wants to protect the data so that it cannot be accessed and transferred to a remote host by malicious code accidentally installed on the training container.
Which action will provide the MOST secure protection?
- A . Encrypt the weights of the CNN model.
- B . Enable network isolation for training jobs.
- C . Remove Amazon S3 access permissions from the SageMaker execution role.
- D . Encrypt the training and validation dataset.