Professional Machine Learning Engineer Exam QuestionsBrowse all questions from this exam

Professional Machine Learning Engineer Exam - Question 37


You are developing models to classify customer support emails. You created models with TensorFlow Estimators using small datasets on your on-premises system, but you now need to train the models using large datasets to ensure high performance. You will port your models to Google Cloud and want to minimize code refactoring and infrastructure overhead for easier migration from on-prem to cloud. What should you do?

Show Answer
Correct Answer: A

To minimize code refactoring and infrastructure overhead when migrating TensorFlow Estimators from on-premises systems to Google Cloud, using AI Platform for distributed training is the best choice. AI Platform provides a managed service for training machine learning models which allows you to leverage the same TensorFlow Estimator code with minimal changes. It handles the distribution of your training jobs across multiple machines automatically, reducing the burden of managing the underlying infrastructure and ensuring efficient training on large datasets.

Discussion

17 comments
Sign in to comment
maartenalexanderOption: A
Jun 22, 2021

A. AI platform provides lower infrastructure overhead and allows you to not have to refactor your code too much (no containerization and such, like in KubeFlow).

q4examOption: B
Sep 22, 2021

I think the answer is either A or B, but personally think it is likely B because dataproc is a common tool box on GCP used for ML while AI platform might require refactoring. However, I dont really know A or B

george_ognyanov
Oct 13, 2021

Another vote for answer A. AI Platform distributed training here. However, I wanted to share my logic why its not B as well. Dataproc is a managed Hadoop and as such needs a processing engine for ML tasks. Most likely Spark and SparkML. Now Spark code is quite different than pure Python and SparkML is even more different than TFcode. I imagine there might me a way to convert TF code to run on SparkML, but this seems a lot of work. And besides the question specifically wants us to minimize refactoring, so there you have it, we can eliminate option B 100%.

NamitSehgalOption: A
Jan 1, 2022

TensorFlow Estimators means require distributed and that is key feature for AI platform or later Vertex AI.

A4MOption: A
Jan 25, 2022

A - better to go with managed service and distributed

David_mlOption: A
May 9, 2022

The answer is A. AI platform also contains kubeflow pipelines. you don't need to set up infrastructure to use it. For D you need to set up a kubernetes cluster engine. The question asks us to minimize infrastructure overheard.

harithacMLOption: A
Jul 13, 2023

A. Use AI Platform for distributed training. : Managed , low infra change migration: yes , although need code refactoring to bigquery sql B. Create a cluster on Dataproc for training. : only cluster ? what about training? C. Create a Managed Instance Group with autoscaling. : Same Q? D. Use Kubeflow Pipelines to train on a Google Kubernetes Engine cluster : only training?

fragkrisOption: A
Dec 5, 2023

I chose A. Even though D is a working option, it requires us to create a GKE cluster, which requires more work.

girguOption: A
May 26, 2024

The most suitable option for minimizing code refactoring and infrastructure overhead while enabling large-scale training on Google Cloud is: A. Use AI Platform for distributed training. * **Simplified Workflow:** AI Platform offers a managed service for training machine learning models. You can train your existing TensorFlow Estimator code with minimal changes, reducing the need for extensive code refactoring. * **Distributed Training:** AI Platform automatically handles distributing your training job across multiple machines, allowing you to leverage the power of Google's cloud infrastructure to train on large datasets efficiently. * **Reduced Infrastructure Overhead:** You don't need to manage the underlying infrastructure (e.g., setting up and maintaining a cluster) yourself. AI Platform takes care of all the infrastructure provisioning and management, minimizing the workload on your team.

JobQOption: A
Dec 20, 2021

I think is A

DHEEPAKOption: D
Jan 16, 2022

I am 100% sure that the answer is D. Kubeflow pipelines were designed keeping: A) Portability. B) Composability. C) Flexibility in mind. This is the pain point that the kubeflow pipelines address

David_ml
May 9, 2022

The answer is A. AI platform also contains kubeflow pipelines. you don't need to set up infrastructure to use it. For D you need to set up a kubernetes cluster engine. The question asks us to minimize infrastructure overheard.

mmona19Option: D
Apr 14, 2022

D- Kubeflow pipelines with Vertex ai provides you ability to reuse existing code using a TF conatiner in a pipeline. it helps automate the process. there is a qwiklab walking through this. A-incorrect, question is asking resuse existing code with minimum changes. distributed deployment does not address that.

David_ml
May 9, 2022

The answer is A. AI platform also contains kubeflow pipelines. you don't need to set up infrastructure to use it. For D you need to set up a kubernetes cluster engine. The question asks us to minimize infrastructure overheard.

Mohamed_MossadOption: A
Jun 13, 2022

using options eliminations answer between A,D will vote for A as it is easier

Mohamed_MossadOption: A
Jun 13, 2022

- using options eliminations answer between A,D will vote for A as it is easier

FatiyOption: A
Feb 28, 2023

Option A is the best choice as AI Platform provides a distributed training framework, enabling you to train large-scale models faster and with less effort

M25Option: A
May 9, 2023

Went with A

Sum_SumOption: A
Nov 15, 2023

A - because it has native support for TF

PhilipKokuOption: A
Jun 6, 2024

A) AI Platform