Pass the Databricks Generative AI Engineer Databricks-Generative-AI-Engineer-Associate Questions and answers with CertsForce

Viewing page 2 out of 2 pages
Viewing questions 11-20 out of questions
Questions # 11:

After changing the response generating LLM in a RAG pipeline from GPT-4 to a model with a shorter context length that the company self-hosts, the Generative AI Engineer is getting the following error:

Question # 11

What TWO solutions should the Generative AI Engineer implement without changing the response generating model? (Choose two.)

Options:

A.

Use a smaller embedding model to generate


B.

Reduce the maximum output tokens of the new model


C.

Decrease the chunk size of embedded documents


D.

Reduce the number of records retrieved from the vector database


E.

Retrain the response generating model using ALiBi


Expert Solution
Questions # 12:

A Generative Al Engineer is developing a RAG application and would like to experiment with different embedding models to improve the application performance.

Which strategy for picking an embedding model should they choose?

Options:

A.

Pick an embedding model trained on related domain knowledge


B.

Pick the most recent and most performant open LLM released at the time


C.

pick the embedding model ranked highest on the Massive Text Embedding Benchmark (MTEB) leaderboard hosted by HuggingFace


D.

Pick an embedding model with multilingual support to support potential multilingual user questions


Expert Solution
Questions # 13:

A Generative Al Engineer is building a RAG application that answers questions about internal documents for the company SnoPen AI.

The source documents may contain a significant amount of irrelevant content, such as advertisements, sports news, or entertainment news, or content about other companies.

Which approach is advisable when building a RAG application to achieve this goal of filtering irrelevant information?

Options:

A.

Keep all articles because the RAG application needs to understand non-company content to avoid answering questions about them.


B.

Include in the system prompt that any information it sees will be about SnoPenAI, even if no data filtering is performed.


C.

Include in the system prompt that the application is not supposed to answer any questions unrelated to SnoPen Al.


D.

Consolidate all SnoPen AI related documents into a single chunk in the vector database.


Expert Solution
Questions # 14:

A Generative Al Engineer is setting up a Databricks Vector Search that will lookup news articles by topic within 10 days of the date specified An example query might be "Tell me about monster truck news around January 5th 1992". They want to do this with the least amount of effort.

How can they set up their Vector Search index to support this use case?

Options:

A.

Split articles by 10 day blocks and return the block closest to the query.


B.

Include metadata columns for article date and topic to support metadata filtering.


C.

pass the query directly to the vector search index and return the best articles.


D.

Create separate indexes by topic and add a classifier model to appropriately pick the best index.


Expert Solution
Questions # 15:

A Generative Al Engineer is helping a cinema extend its website's chat bot to be able to respond to questions about specific showtimes for movies currently playing at their local theater. They already have the location of the user provided by location services to their agent, and a Delta table which is continually updated with the latest showtime information by location. They want to implement this new capability In their RAG application.

Which option will do this with the least effort and in the most performant way?

Options:

A.

Create a Feature Serving Endpoint from a FeatureSpec that references an online store synced from the Delta table. Query the Feature Serving Endpoint as part of the agent logic / tool implementation.


B.

Query the Delta table directly via a SQL query constructed from the user's input using a text-to-SQL LLM in the agent logic / tool


C.

implementation. Write the Delta table contents to a text column.then embed those texts using an embedding model and store these in the vector index Look

up the information based on the embedding as part of the agent logic / tool implementation.


D.

Set up a task in Databricks Workflows to write the information in the Delta table periodically to an external database such as MySQL and query the information from there as part of the agent logic / tool implementation.


Expert Solution
Questions # 16:

A Generative Al Engineer needs to design an LLM pipeline to conduct multi-stage reasoning that leverages external tools. To be effective at this, the LLM will need to plan and adapt actions while performing complex reasoning tasks.

Which approach will do this?

Options:

A.

Tram the LLM to generate a single, comprehensive response without interacting with any external tools, relying solely on its pre-trained knowledge.


B.

Implement a framework like ReAct which allows the LLM to generate reasoning traces and perform task-specific actions that leverage external tools if necessary.


C.

Encourage the LLM to make multiple API calls in sequence without planning or structuring the calls, allowing the LLM to decide when and how to use external tools spontaneously.


D.

Use a Chain-of-Thought (CoT) prompting technique to guide the LLM through a series of reasoning steps, then manually input the results from external tools for the final answer.


Expert Solution
Questions # 17:

A Generative Al Engineer is tasked with developing a RAG application that will help a small internal group of experts at their company answer specific questions, augmented by an internal knowledge base. They want the best possible quality in the answers, and neither latency nor throughput is a huge concern given that the user group is small and they’re willing to wait for the best answer. The topics are sensitive in nature and the data is highly confidential and so, due to regulatory requirements, none of the information is allowed to be transmitted to third parties.

Which model meets all the Generative Al Engineer’s needs in this situation?

Options:

A.

Dolly 1.5B


B.

OpenAI GPT-4


C.

BGE-large


D.

Llama2-70B


Expert Solution
Questions # 18:

A Generative AI Engineer is tasked with deploying an application that takes advantage of a custom MLflow Pyfunc model to return some interim results.

How should they configure the endpoint to pass the secrets and credentials?

Options:

A.

Use spark.conf.set ()


B.

Pass variables using the Databricks Feature Store API


C.

Add credentials using environment variables


D.

Pass the secrets in plain text


Expert Solution
Viewing page 2 out of 2 pages
Viewing questions 11-20 out of questions