Spring Sale Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: simple70

Pass the Google Machine Learning Engineer Professional-Machine-Learning-Engineer Questions and answers with CertsForce

Viewing page 7 out of 9 pages
Viewing questions 61-70 out of questions
Questions # 61:

You recently developed a deep learning model using Keras, and now you are experimenting with different training strategies. First, you trained the model using a single GPU, but the training process was too slow. Next, you distributed the training across 4 GPUs using tf.distribute.MirroredStrategy (with no other changes), but you did not observe a decrease in training time. What should you do?

Options:

A.

Distribute the dataset with tf.distribute.Strategy.experimental_distribute_dataset


B.

Create a custom training loop.


C.

Use a TPU with tf.distribute.TPUStrategy.


D.

Increase the batch size.


Expert Solution
Questions # 62:

You are responsible for building a unified analytics environment across a variety of on-premises data marts. Your company is experiencing data quality and security challenges when integrating data across the servers, caused by the use of a wide range of disconnected tools and temporary solutions. You need a fully managed, cloud-native data integration service that will lower the total cost of work and reduce repetitive work. Some members on your team prefer a codeless interface for building Extract, Transform, Load (ETL) process. Which service should you use?

Options:

A.

Dataflow


B.

Dataprep


C.

Apache Flink


D.

Cloud Data Fusion


Expert Solution
Questions # 63:

You recently built the first version of an image segmentation model for a self-driving car. After deploying the model, you observe a decrease in the area under the curve (AUC) metric. When analyzing the video recordings, you also discover that the model fails in highly congested traffic but works as expected when there is less traffic. What is the most likely reason for this result?

Options:

A.

The model is overfitting in areas with less traffic and underfitting in areas with more traffic.


B.

AUC is not the correct metric to evaluate this classification model.


C.

Too much data representing congested areas was used for model training.


D.

Gradients become small and vanish while backpropagating from the output to input nodes.


Expert Solution
Questions # 64:

You are an ML engineer in the contact center of a large enterprise. You need to build a sentiment analysis tool that predicts customer sentiment from recorded phone conversations. You need to identify the best approach to building a model while ensuring that the gender, age, and cultural differences of the customers who called the contact center do not impact any stage of the model development pipeline and results. What should you do?

Options:

A.

Extract sentiment directly from the voice recordings


B.

Convert the speech to text and build a model based on the words


C.

Convert the speech to text and extract sentiments based on the sentences


D.

Convert the speech to text and extract sentiment using syntactical analysis


Expert Solution
Questions # 65:

Your company ' s business stakeholders want to understand the factors driving customer churn to inform their business strategy. You need to build a customer churn prediction model that prioritizes simple interpretability of your model ' s results. You need to choose the ML framework and modeling technique that will explain which features led to the prediction. What should you do?

Options:

A.

Build a TensorFlow deep neural network (DNN) model, and use SHAP values for feature importance analysis.


B.

Build a PyTorch long short-term memory (LSTM) network, and use attention mechanisms for interpretability.


C.

Build a logistic regression model in scikit-learn, and interpret the model ' s output coefficients to understand feature impact.


D.

Build a linear regression model in scikit-learn, and interpret the model ' s standardized coefficients to understand feature impact.


Expert Solution
Questions # 66:

You have created a Vertex Al pipeline that includes two steps. The first step preprocesses 10 TB data completes in about 1 hour, and saves the result in a Cloud Storage bucket The second step uses the processed data to train a model You need to update the model ' s code to allow you to test different algorithms You want to reduce pipeline execution time and cost, while also minimizing pipeline changes What should you do?

Options:

A.

Add a pipeline parameter and an additional pipeline step Depending on the parameter value the pipeline step conducts or skips data preprocessing and starts model training.


B.

Create another pipeline without the preprocessing step, and hardcode the preprocessed Cloud Storage file location for model training.


C.

Configure a machine with more CPU and RAM from the compute-optimized machine family for the data preprocessing step.


D.

Enable caching for the pipeline job. and disable caching for the model training step.


Expert Solution
Questions # 67:

You are analyzing customer data for a healthcare organization that is stored in Cloud Storage. The data contains personally identifiable information (PII) You need to perform data exploration and preprocessing while ensuring the security and privacy of sensitive fields What should you do?

Options:

A.

Use the Cloud Data Loss Prevention (DLP) API to de-identify the PI! before performing data exploration and preprocessing.


B.

Use customer-managed encryption keys (CMEK) to encrypt the Pll data at rest and decrypt the Pll data during data exploration and preprocessing.


C.

Use a VM inside a VPC Service Controls security perimeter to perform data exploration and preprocessing.


D.

Use Google-managed encryption keys to encrypt the Pll data at rest, and decrypt the Pll data during data exploration and preprocessing.


Expert Solution
Questions # 68:

You have successfully deployed to production a large and complex TensorFlow model trained on tabular data. You want to predict the lifetime value (LTV) field for each subscription stored in the BigQuery table named subscription. subscriptionPurchase in the project named my-fortune500-company-project.

You have organized all your training code, from preprocessing data from the BigQuery table up to deploying the validated model to the Vertex AI endpoint, into a TensorFlow Extended (TFX) pipeline. You want to prevent prediction drift, i.e., a situation when a feature data distribution in production changes significantly over time. What should you do?

Options:

A.

Implement continuous retraining of the model daily using Vertex AI Pipelines.


B.

Add a model monitoring job where 10% of incoming predictions are sampled 24 hours.


C.

Add a model monitoring job where 90% of incoming predictions are sampled 24 hours.


D.

Add a model monitoring job where 10% of incoming predictions are sampled every hour.


Expert Solution
Questions # 69:

You need to train a computer vision model that predicts the type of government ID present in a given image using a GPU-powered virtual machine on Compute Engine. You use the following parameters:

• Optimizer: SGD

• Image shape = 224x224

• Batch size = 64

• Epochs = 10

• Verbose = 2

During training you encounter the following error: ResourceExhaustedError: out of Memory (oom) when allocating tensor. What should you do?

Options:

A.

Change the optimizer


B.

Reduce the batch size


C.

Change the learning rate


D.

Reduce the image shape


Expert Solution
Questions # 70:

You are building a linear model with over 100 input features, all with values between -1 and 1. You suspect that many features are non-informative. You want to remove the non-informative features from your model while keeping the informative ones in their original form. Which technique should you use?

Options:

A.

Use Principal Component Analysis to eliminate the least informative features.


B.

Use L1 regularization to reduce the coefficients of uninformative features to 0.


C.

After building your model, use Shapley values to determine which features are the most informative.


D.

Use an iterative dropout technique to identify which features do not degrade the model when removed.


Expert Solution
Viewing page 7 out of 9 pages
Viewing questions 61-70 out of questions