Spring Sale Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: simple70

Pass the Google Google Cloud Certified Professional-Data-Engineer Questions and answers with CertsForce

Viewing page 4 out of 8 pages
Viewing questions 31-40 out of questions
Questions # 31:

You are deploying a new storage system for your mobile application, which is a media streaming service. You decide the best fit is Google Cloud Datastore. You have entities with multiple properties, some of which can take on multiple values. For example, in the entity ‘Movie’ the property ‘actors’ and the property ‘tags’ have multiple values but the property ‘date released’ does not. A typical query would ask for all movies with actor=<actorname> ordered by date_released or all movies with tag=Comedy ordered by date_released. How should you avoid a combinatorial explosion in the number of indexes?

Question # 31

Options:

A.

Option A


B.

Option B.


C.

Option C


D.

Option D


Expert Solution
Questions # 32:

You work for an economic consulting firm that helps companies identify economic trends as they happen. As part of your analysis, you use Google BigQuery to correlate customer data with the average prices of the 100 most common goods sold, including bread, gasoline, milk, and others. The average prices of these goods are updated every 30 minutes. You want to make sure this data stays up to date so you can combine it with other data in BigQuery as cheaply as possible. What should you do?

Options:

A.

Load the data every 30 minutes into a new partitioned table in BigQuery.


B.

Store and update the data in a regional Google Cloud Storage bucket and create a federated data source in BigQuery


C.

Store the data in Google Cloud Datastore. Use Google Cloud Dataflow to query BigQuery and combine the data programmatically with the data stored in Cloud Datastore


D.

Store the data in a file in a regional Google Cloud Storage bucket. Use Cloud Dataflow to query BigQuery and combine the data programmatically with the data stored in Google Cloud Storage.


Expert Solution
Questions # 33:

Your company has recently grown rapidly and now ingesting data at a significantly higher rate than it was previously. You manage the daily batch MapReduce analytics jobs in Apache Hadoop. However, the recent increase in data has meant the batch jobs are falling behind. You were asked to recommend ways the development team could increase the responsiveness of the analytics without increasing costs. What should you recommend they do?

Options:

A.

Rewrite the job in Pig.


B.

Rewrite the job in Apache Spark.


C.

Increase the size of the Hadoop cluster.


D.

Decrease the size of the Hadoop cluster but also rewrite the job in Hive.


Expert Solution
Questions # 34:

You work for a manufacturing plant that batches application log files together into a single log file once a day at 2:00 AM. You have written a Google Cloud Dataflow job to process that log file. You need to make sure the log file in processed once per day as inexpensively as possible. What should you do?

Options:

A.

Change the processing job to use Google Cloud Dataproc instead.


B.

Manually start the Cloud Dataflow job each morning when you get into the office.


C.

Create a cron job with Google App Engine Cron Service to run the Cloud Dataflow job.


D.

Configure the Cloud Dataflow job as a streaming job so that it processes the log data immediately.


Expert Solution
Questions # 35:

You are designing the database schema for a machine learning-based food ordering service that will predict what users want to eat. Here is some of the information you need to store:

The user profile: What the user likes and doesn’t like to eat

The user account information: Name, address, preferred meal times

The order information: When orders are made, from where, to whom

The database will be used to store all the transactional data of the product. You want to optimize the data schema. Which Google Cloud Platform product should you use?

Options:

A.

BigQuery


B.

Cloud SQL


C.

Cloud Bigtable


D.

Cloud Datastore


Expert Solution
Questions # 36:

Which of the following job types are supported by Cloud Dataproc (select 3 answers)?

Options:

A.

Hive


B.

Pig


C.

YARN


D.

Spark


Expert Solution
Questions # 37:

What are two of the benefits of using denormalized data structures in BigQuery?

Options:

A.

Reduces the amount of data processed, reduces the amount of storage required


B.

Increases query speed, makes queries simpler


C.

Reduces the amount of storage required, increases query speed


D.

Reduces the amount of data processed, increases query speed


Expert Solution
Questions # 38:

Suppose you have a dataset of images that are each labeled as to whether or not they contain a human face. To create a neural network that recognizes human faces in images using this labeled dataset, what approach would likely be the most effective?

Options:

A.

Use K-means Clustering to detect faces in the pixels.


B.

Use feature engineering to add features for eyes, noses, and mouths to the input data.


C.

Use deep learning by creating a neural network with multiple hidden layers to automatically detect features of faces.


D.

Build a neural network with an input layer of pixels, a hidden layer, and an output layer with two categories.


Expert Solution
Questions # 39:

Which of the following is NOT one of the three main types of triggers that Dataflow supports?

Options:

A.

Trigger based on element size in bytes


B.

Trigger that is a combination of other triggers


C.

Trigger based on element count


D.

Trigger based on time


Expert Solution
Questions # 40:

You work for a bank. You have a labelled dataset that contains information on already granted loan application and whether these applications have been defaulted. You have been asked to train a model to predict default rates for credit applicants.

What should you do?

Options:

A.

Increase the size of the dataset by collecting additional data.


B.

Train a linear regression to predict a credit default risk score.


C.

Remove the bias from the data and collect applications that have been declined loans.


D.

Match loan applicants with their social profiles to enable feature engineering.


Expert Solution
Viewing page 4 out of 8 pages
Viewing questions 31-40 out of questions