AWS Certified Machine Learning Engineer - Associate MLA-C01 Exam QuestionsBrowse all questions from this exam

AWS Certified Machine Learning Engineer - Associate MLA-C01 Exam - Question 91


A company runs Amazon SageMaker ML models that use accelerated instances. The models require real-time responses. Each model has different scaling requirements. The company must not allow a cold start for the models.

Which solution will meet these requirements?

Show Answer
Correct Answer:

Discussion

2 comments
Sign in to comment
eesaOption: C
Mar 21, 2025

✅ Explanation: Requirements Recap: Real-time inference: Needs low-latency predictions. Accelerated instances: Likely GPU-backed, costly to scale inefficiently. No cold starts: Endpoints must always be warm and responsive. Each model has different scaling needs: Must support independent scaling of each model. ✅ Why Option C is correct: Inference components are a new SageMaker feature that allow: Hosting multiple models on a single endpoint. Independent scaling of each model (component). Avoiding cold starts via minimum number of copies. Setting min invocations or min replicas ≥ 1 keeps the model always warm, eliminating cold starts. This solution meets all requirements efficiently.

ygn4eiOption: A
Mar 20, 2025

this is correct