Professional Machine Learning Engineer Exam QuestionsBrowse all questions from this exam

Professional Machine Learning Engineer Exam - Question 83


You need to design an architecture that serves asynchronous predictions to determine whether a particular mission-critical machine part will fail. Your system collects data from multiple sensors from the machine. You want to build a model that will predict a failure in the next N minutes, given the average of each sensor’s data from the past 12 hours. How should you design the architecture?

Show Answer
Correct Answer: B

The architecture should be designed to handle real-time processing and prediction since the requirement is to predict whether a machine part will fail in the next N minutes, given the average of each sensor’s data from the past 12 hours. This requires handling continuous streams of sensor data and making timely predictions. Using Pub/Sub for event ingestion and Dataflow for real-time processing, followed by invoking the model for predictions and sending the results to another Pub/Sub topic for downstream consumption, ensures that predictions can be made and consumed with minimal latency. This setup is ideal for mission-critical applications where timely and continuous predictions are necessary. This makes option B the best fit for the given requirements.

Discussion

17 comments
Sign in to comment
hiromiOption: B
Dec 18, 2022

B "Predictive maintenance: asynchronously predicting whether a particular machine part will fail in the next N minutes, given the averages of the sensor's data in the past 30 minutes." https://cloud.google.com/architecture/minimizing-predictive-serving-latency-in-machine-learning#offline_batch_prediction

hiromi
Dec 23, 2022

- https://cloud.google.com/architecture/minimizing-predictive-serving-latency-in-machine-learning#online_real-time_prediction

John_PongthornOption: B
Jan 27, 2023

B is most likely . if you search asynchronous on this page. it appears in the question wants to focus on online prediction with asynchronous mode. https://cloud.google.com/architecture/minimizing-predictive-serving-latency-in-machine-learning#online_real-time_prediction and the question is the same as what has been explained in this section obviously. it is as below. Predictive maintenance: asynchronously predicting whether a particular machine part will fail in the next N minutes, given the averages of the sensor's data in the past 30 minutes. afte that, you can take a closer look at figure3 and read what it try to describle C and D it is the offline solution but you opt to use different tools. https://cloud.google.com/architecture/minimizing-predictive-serving-latency-in-machine-learning#offline_batch_prediction

tavva_prudhviOption: B
Mar 22, 2023

Option C may not be the best choice for this use case because it involves using a batch prediction job in Vertex AI to perform scoring on preprocessed data. Batch prediction jobs are more suitable for scenarios where data is processed in batches, and results can be generated over a longer period, such as daily or weekly. In this use case, the requirement is to predict whether a machine part will fail in the next N minutes, given the average of each sensor's data from the past 12 hours. Therefore, real-time processing and prediction are necessary. Batch prediction jobs are not designed for real-time processing, and there may be a delay in receiving the predictions. Option B, on the other hand, is designed for real-time processing and prediction. The Pub/Sub and Dataflow components allow for real-time processing of incoming sensor data, and the trained ML model can be invoked for prediction in real-time. This makes it ideal for mission-critical applications where timely predictions are essential.

JamesDoeOption: B
Mar 28, 2023

B. Online prediction, and need decoupling with Pub/Sub to make it asynchronous. Option A is synchronous.

rosenr0Option: D
May 28, 2023

D. I think we have to query data from the past 12 hours for the prediction, and that's the reason for exporting the data to Cloud Storage. Also, the predictions don't have to be real time.

pawan94Option: B
Jan 9, 2024

Here you go to the answer provided by google itself. I don't understand why would people use batch prediction when they its sensor data and online prediction is as well asynchronous. https://cloud.google.com/architecture/minimizing-predictive-serving-latency-in-machine-learning#offline_batch_prediction:~:text=Predictive%20maintenance%3A%20asynchronously%20predicting%20whether%20a%20particular%20machine%20part%20will%20fail%20in%20the%20next%20N%20minutes%2C%20given%20the%20averages%20of%20the%20sensor%27s%20data%20in%20the%20past%2030%20minutes.

seifouOption: C
Dec 13, 2022

ref : https://cloud.google.com/architecture/minimizing-predictive-serving-latency-in-machine-learning#offline_batch_prediction

ares81Option: C
Dec 14, 2022

C, for me.

mil_spyroOption: B
Dec 17, 2022

Answer is B. https://cloud.google.com/architecture/minimizing-predictive-serving-latency-in-machine-learning#handling_dynamic_real-time_features

John_PongthornOption: C
Jan 27, 2023

Asycnchromoue preciction = Batch prediction https://cloud.google.com/architecture/minimizing-predictive-serving-latency-in-machine-learning#offline_batch_prediction

John_Pongthorn
Jan 27, 2023

Asynchronous prediction = Batch prediction, It is incorrect because I am reckless to read this article, Admin can delete my shitty comment above. I was mistaken

enghabethOption: B
Feb 9, 2023

if you have sensors inyour architecture.. you need pub/sub...

tavva_prudhviOption: B
Mar 22, 2023

Its B, This architecture leverages the strengths of Pub/Sub, Dataflow, and Vertex AI. The system collects data from multiple sensors, which sends events to Pub/Sub. Pub/Sub can handle the high volume of incoming data and can buffer messages to prevent data loss. A Dataflow stream processing pipeline can consume the events in real-time and perform feature engineering and data preprocessing before invoking the trained ML model for prediction. The predictions are then sent to another Pub/Sub topic, where they can be consumed by a downstream system for monitoring. This architecture is highly scalable, resilient, and efficient, as it can handle large volumes of data and perform real-time processing and prediction. It also separates concerns by using a separate pipeline for data processing and another for prediction, making it easier to maintain and modify the system.

M25Option: B
May 9, 2023

Went with B

vale_76_na_xxxOption: C
Dec 19, 2023

it refers to asincronou prediction I' go with C

Werner123Option: B
Feb 29, 2024

Needs to be real time not batch. The data needs to be processed as a stream since multiple sensors are used. pawan94 is right. https://cloud.google.com/architecture/minimizing-predictive-serving-latency-in-machine-learning#online_real-time_prediction

andreabrunelliOption: C
Apr 24, 2024

The simplest solution that can support an eventual batch prediction (triggered by pub/sub) even the semi-real time prediction.

PhilipKokuOption: B
Jun 7, 2024

B) Pub/Sub & DataFlow