Spring Sale Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: simple70

Pass the Google Machine Learning Engineer Professional-Machine-Learning-Engineer Questions and answers with CertsForce

Viewing page 6 out of 9 pages
Viewing questions 51-60 out of questions
Questions # 51:

You are developing a model to identify traffic signs in images extracted from videos taken from the dashboard of a vehicle. You have a dataset of 100 000 images that were cropped to show one out of ten different traffic signs. The images have been labeled accordingly for model training and are stored in a Cloud Storage bucket You need to be able to tune the model during each training run. How should you train the model?

Options:

A.

Train a model for object detection by using Vertex Al AutoML.


B.

Train a model for image classification by using Vertex Al AutoML.


C.

Develop the model training code for object detection and tram a model by using Vertex Al custom training.


D.

Develop the model training code for image classification and train a model by using Vertex Al custom training.


Expert Solution
Questions # 52:

You are developing ML models with Al Platform for image segmentation on CT scans. You frequently update your model architectures based on the newest available research papers, and have to rerun training on the same dataset to benchmark their performance. You want to minimize computation costs and manual intervention while having version control for your code. What should you do?

Options:

A.

Use Cloud Functions to identify changes to your code in Cloud Storage and trigger a retraining job


B.

Use the gcloud command-line tool to submit training jobs on Al Platform when you update your code


C.

Use Cloud Build linked with Cloud Source Repositories to trigger retraining when new code is pushed to the repository


D.

Create an automated workflow in Cloud Composer that runs daily and looks for changes in code in Cloud Storage using a sensor.


Expert Solution
Questions # 53:

You manage a team of data scientists who use a cloud-based backend system to submit training jobs. This system has become very difficult to administer, and you want to use a managed service instead. The data scientists you work with use many different frameworks, including Keras, PyTorch, theano, scikit-learn, and custom libraries. What should you do?

Options:

A.

Use the Vertex AI Training to submit training jobs using any framework.


B.

Configure Kubeflow to run on Google Kubernetes Engine and submit training jobs through TFJob.


C.

Create a library of VM images on Compute Engine, and publish these images on a centralized repository.


D.

Set up Slurm workload manager to receive jobs that can be scheduled to run on your cloud infrastructure.


Expert Solution
Questions # 54:

You recently deployed a model lo a Vertex Al endpoint and set up online serving in Vertex Al Feature Store. You have configured a daily batch ingestion job to update your featurestore During the batch ingestion jobs you discover that CPU utilization is high in your featurestores online serving nodes and that feature retrieval latency is high. You need to improve online serving performance during the daily batch ingestion. What should you do?

Options:

A.

Schedule an increase in the number of online serving nodes in your featurestore prior to the batch ingestion jobs.


B.

Enable autoscaling of the online serving nodes in your featurestore


C.

Enable autoscaling for the prediction nodes of your DeployedModel in the Vertex Al endpoint.


D.

Increase the worker counts in the importFeaturevalues request of your batch ingestion job.


Expert Solution
Questions # 55:

You are building an ML model to predict trends in the stock market based on a wide range of factors. While exploring the data, you notice that some features have a large range. You want to ensure that the features with the largest magnitude don’t overfit the model. What should you do?

Options:

A.

Standardize the data by transforming it with a logarithmic function.


B.

Apply a principal component analysis (PCA) to minimize the effect of any particular feature.


C.

Use a binning strategy to replace the magnitude of each feature with the appropriate bin number.


D.

Normalize the data by scaling it to have values between 0 and 1.


Expert Solution
Questions # 56:

One of your models is trained using data provided by a third-party data broker. The data broker does not reliably notify you of formatting changes in the data. You want to make your model training pipeline more robust to issues like this. What should you do?

Options:

A.

Use TensorFlow Data Validation to detect and flag schema anomalies.


B.

Use TensorFlow Transform to create a preprocessing component that will normalize data to the expected distribution, and replace values that don’t match the schema with 0.


C.

Use tf.math to analyze the data, compute summary statistics, and flag statistical anomalies.


D.

Use custom TensorFlow functions at the start of your model training to detect and flag known formatting errors.


Expert Solution
Questions # 57:

You are an ML engineer on an agricultural research team working on a crop disease detection tool to detect leaf rust spots in images of crops to determine the presence of a disease. These spots, which can vary in shape and size, are correlated to the severity of the disease. You want to develop a solution that predicts the presence and severity of the disease with high accuracy. What should you do?

Options:

A.

Create an object detection model that can localize the rust spots.


B.

Develop an image segmentation ML model to locate the boundaries of the rust spots.


C.

Develop a template matching algorithm using traditional computer vision libraries.


D.

Develop an image classification ML model to predict the presence of the disease.


Expert Solution
Questions # 58:

Your company manages an ecommerce platform and has a large dataset of customer reviews. Each review has a positive, negative, or neutral label. You need to quickly prototype a sentiment analysis model that accurately predicts the sentiment labels of new customer reviews while minimizing time and cost. What should you do?

Options:

A.

Train a sentiment analysis model by using a BERT-based model, and fine-tune the model by using domain-specific customer reviews.


B.

Use the Natural Language API for real-time sentiment analysis.


C.

Use AutoML to train a multi-class classification model that predicts sentiment labels based on the training data.


D.

Use the Vertex AI Text embeddings API to vectorize the text, and train a regression model by using AutoML to predict sentiment scores.


Expert Solution
Questions # 59:

You work for a large retailer and you need to build a model to predict customer churn. The company has a dataset of historical customer data, including customer demographics, purchase history, and website activity. You need to create the model in BigQuery ML and thoroughly evaluate its performance. What should you do?

Options:

A.

Create a linear regression model in BigQuery ML and register the model in Vertex Al Model Registry Evaluate the model performance in Vertex Al.


B.

Create a logistic regression model in BigQuery ML and register the model in Vertex Al Model Registry. Evaluate the model performance in Vertex Al.


C.

Create a linear regression model in BigQuery ML Use the ml. evaluate function to evaluate the model performance.


D.

Create a logistic regression model in BigQuery ML Use the ml.confusion_matrix function to evaluate the model performance.


Expert Solution
Questions # 60:

You are building a model to predict daily temperatures. You split the data randomly and then transformed the training and test datasets. Temperature data for model training is uploaded hourly. During testing, your model performed with 97% accuracy; however, after deploying to production, the model ' s accuracy dropped to 66%. How can you make your production model more accurate?

Options:

A.

Normalize the data for the training, and test datasets as two separate steps.


B.

Split the training and test data based on time rather than a random split to avoid leakage


C.

Add more data to your test set to ensure that you have a fair distribution and sample for testing


D.

Apply data transformations before splitting, and cross-validate to make sure that the transformations are applied to both the training and test sets.


Expert Solution
Viewing page 6 out of 9 pages
Viewing questions 51-60 out of questions