Spring Sale Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: simple70

Pass the Google Cloud DevOps Engineer Professional-Cloud-DevOps-Engineer Questions and answers with CertsForce

Viewing page 1 out of 6 pages
Viewing questions 1-10 out of questions
Questions # 1:

You are running an application on Compute Engine and collecting logs through Stackdriver. You discover that some personally identifiable information (Pll) is leaking into certain log entry fields. All Pll entries begin with the text userinfo. You want to capture these log entries in a secure location for later review and prevent them from leaking to Stackdriver Logging. What should you do?

Options:

A.

Create a basic log filter matching userinfo, and then configure a log export in the Stackdriver console with Cloud Storage as a sink.


B.

Use a Fluentd filter plugin with the Stackdriver Agent to remove log entries containing userinfo, and then copy the entries to a Cloud Storage bucket.


C.

Create an advanced log filter matching userinfo, configure a log export in the Stackdriver console with Cloud Storage as a sink, and then configure a tog exclusion with userinfo as a filter.


D.

Use a Fluentd filter plugin with the Stackdriver Agent to remove log entries containing userinfo, create an advanced log filter matching userinfo, and then configure a log export in the Stackdriver console with Cloud Storage as a sink.


Expert Solution
Questions # 2:

You are using Terraform to manage infrastructure as code within a Cl/CD pipeline You notice that multiple copies of the entire infrastructure stack exist in your Google Cloud project, and a new copy is created each time a change to the existing infrastructure is made You need to optimize your cloud spend by ensuring that only a single instance of your infrastructure stack exists at a time. You want to follow Google-recommended practices What should you do?

Options:

A.

Create a new pipeline to delete old infrastructure stacks when they are no longer needed


B.

Confirm that the pipeline is storing and retrieving the terraform. if state file from Cloud Storage with the Terraform gcs backend


C.

Verify that the pipeline is storing and retrieving the terrafom.tfstat* file from a source control


D.

Update the pipeline to remove any existing infrastructure before you apply the latest configuration


Expert Solution
Questions # 3:

Your company follows Site Reliability Engineering principles. You are writing a postmortem for an incident, triggered by a software change, that severely affected users. You want to prevent severe incidents from happening in the future. What should you do?

Options:

A.

Identify engineers responsible for the incident and escalate to their senior management.


B.

Ensure that test cases that catch errors of this type are run successfully before new software releases.


C.

Follow up with the employees who reviewed the changes and prescribe practices they should follow in the future.


D.

Design a policy that will require on-call teams to immediately call engineers and management to discuss a plan of action if an incident occurs.


Expert Solution
Questions # 4:

Your team is designing a new application for deployment into Google Kubernetes Engine (GKE). You need to set up monitoring to collect and aggregate various application-level metrics in a centralized location. You want to use Google Cloud Platform services while minimizing the amount of work required to set up monitoring. What should you do?

Options:

A.

Publish various metrics from the application directly to the Slackdriver Monitoring API, and then observe these custom metrics in Stackdriver.


B.

Install the Cloud Pub/Sub client libraries, push various metrics from the application to various topics, and then observe the aggregated metrics in Stackdriver.


C.

Install the OpenTelemetry client libraries in the application, configure Stackdriver as the export destination for the metrics, and then observe the application's metrics in Stackdriver.


D.

Emit all metrics in the form of application-specific log messages, pass these messages from the containers to the Stackdriver logging collector, and then observe metrics in Stackdriver.


Expert Solution
Questions # 5:

You need to enforce several constraint templates across your Google Kubernetes Engine (GKE) clusters. The constraints include policy parameters, such as restricting the Kubernetes API. You must ensure that the policy parameters are stored in a GitHub repository and automatically applied when changes occur. What should you do?  

Options:

A.

Set up a GitHub action to trigger Cloud Build when there is a parameter change. In Cloud Build, run a gcloud CLI command to apply the change.


B.

When there is a change in GitHub, use a webhook to send a request to Cloud Service Mesh, and apply the change.


C.

Configure Config Sync with the GitHub repository. When there is a change in the repository, use Config Sync to apply the change.


D.

Configure Config Connector with the GitHub repository. When there is a change in the repository, use Config Connector to apply the change.


Expert Solution
Questions # 6:

You are the Site Reliability Engineer responsible for managing your company's data services and products. You regularly navigate operational challenges, such as unpredictable data volume and high cost, with your company's data ingestion processes. You recently learned that a new data ingestion product will be developed in Google Cloud. You need to collaborate with the product development team to provide operational input on the new product. What should you do?

Options:

A.

Deploy the prototype product in a test environment, run a load test, and share the results with the product development team.


B.

When the initial product version passes the quality assurance phase and compliance assessments, deploy the product to a staging environment. Share error logs and performancemetrics with the product development team.


C.

When the new product is used by at least one internal customer in production, share error logs and monitoring metrics with the product development team.


D.

Review the design of the product with the product development team to provide feedback early in the design phase.


Expert Solution
Questions # 7:

You are running a web application that connects to an AlloyDB cluster by using a private IP address in your default VPC. You need to run a database schema migration in your CI/CD pipeline by using Cloud Build before deploying a new version of your application. You want to follow Google-recommended security practices. What should you do?  

Options:

A.

Set up a Cloud Build private pool to access the database through a static external IP address. Configure the database to only allow connections from this IP address. Execute the schema migration script in the private pool.


B.

Create a service account that has permission to access the database. Configure Cloud Build to use this service account and execute the schema migration script in a private pool.


C.

Add the database username and encrypted password to the application configuration file. Use these credentials in Cloud Build to execute the schema migration script.


D.

Add the database username and password to Secret Manager. When running the schema migration script, retrieve the username and password from Secret Manager.


Expert Solution
Questions # 8:

You have a CI/CD pipeline that uses Cloud Build to build new Docker images and push them to Docker Hub. You use Git for code versioning. After making a change in the Cloud Build YAML configuration, you notice that no new artifacts are being built by the pipeline. You need to resolve the issue following Site Reliability Engineering practices. What should you do?

Options:

A.

Disable the CI pipeline and revert to manually building and pushing the artifacts.


B.

Change the CI pipeline to push the artifacts to Container Registry instead of Docker Hub.


C.

Upload the configuration YAML file to Cloud Storage and use Error Reporting to identify and fix the issue.


D.

Run a Git compare between the previous and current Cloud Build Configuration files to find and fix the bug.


Expert Solution
Questions # 9:

Your organization is using Helm to package containerized applications Your applications reference both public and private charts Your security team flagged that using a public Helm repository as a dependency is a risk You want to manage all charts uniformly, with native access control and VPC Service Controls What should you do?

Options:

A.

Store public and private charts in OCI format by using Artifact Registry


B.

Store public and private charts by using GitHub Enterprise with Google Workspace as the identity provider


C.

Store public and private charts by using Git repository Configure Cloud Build to synchronize contents of the repository into a Cloud Storage bucket Connect Helm to the bucket by using https: // [bucket] .srorage.googleapis.com/ [holnchart] as the Helm repository


D.

Configure a Helm chart repository server to run in Google Kubernetes Engine (GKE) with Cloud Storage bucket as the storage backend


Expert Solution
Questions # 10:

You support a large service with a well-defined Service Level Objective (SLO). The development team deploys new releases of the service multiple times a week. If a major incident causes the service to miss its SLO, you want the development team to shift its focus from working on features to improving service reliability. What should you do before a major incident occurs?

Options:

A.

Develop an appropriate error budget policy in cooperation with all service stakeholders.


B.

Negotiate with the product team to always prioritize service reliability over releasing new features.


C.

Negotiate with the development team to reduce the release frequency to no more than once a week.


D.

Add a plugin to your Jenkins pipeline that prevents new releases whenever your service is out of SLO.


Expert Solution
Viewing page 1 out of 6 pages
Viewing questions 1-10 out of questions