Spring Sale Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: pass65

Databricks-Generative-AI-Engineer-Associate Databricks Certified Generative AI Engineer Associate Questions and Answers

Questions 4

A Generative Al Engineer is building an LLM-based application that has an

important transcription (speech-to-text) task. Speed is essential for the success of the application

Which open Generative Al models should be used?

Options:

A.

L!ama-2-70b-chat-hf

B.

MPT-30B-lnstruct

C.

DBRX

D.

whisper-large-v3 (1.6B)

Buy Now
Questions 5

A Generative AI Engineer is experimenting with using parameters to configure an agent in Mosaic Agent Framework. However, they are struggling to get the agent to respond with relevant information with this configuration:

config = {"prompt_template": "You are a trivia bot. Generate a question based on the user's input: {user_input}", "input_vars": ["user_input"], "parameters": {"temperature": 0.01, "max_tokens": 500}}

Which error is causing the problem?

Options:

A.

The prompt does not parse the user's input vars

B.

The prompt does not set the retriever schema

C.

The prompt does not list available agents for the LLM to call

D.

The prompt is not wrapped in ChatModel

Buy Now
Questions 6

A Generative Al Engineer is building a RAG application that answers questions about internal documents for the company SnoPen AI.

The source documents may contain a significant amount of irrelevant content, such as advertisements, sports news, or entertainment news, or content about other companies.

Which approach is advisable when building a RAG application to achieve this goal of filtering irrelevant information?

Options:

A.

Keep all articles because the RAG application needs to understand non-company content to avoid answering questions about them.

B.

Include in the system prompt that any information it sees will be about SnoPenAI, even if no data filtering is performed.

C.

Include in the system prompt that the application is not supposed to answer any questions unrelated to SnoPen Al.

D.

Consolidate all SnoPen AI related documents into a single chunk in the vector database.

Buy Now
Questions 7

Which indicator should be considered to evaluate the safety of the LLM outputs when qualitatively assessing LLM responses for a translation use case?

Options:

A.

The ability to generate responses in code

B.

The similarity to the previous language

C.

The latency of the response and the length of text generated

D.

The accuracy and relevance of the responses

Buy Now
Questions 8

A Generative Al Engineer is tasked with developing an application that is based on an open source large language model (LLM). They need a foundation LLM with a large context window.

Which model fits this need?

Options:

A.

DistilBERT

B.

MPT-30B

C.

Llama2-70B

D.

DBRX

Buy Now
Questions 9

A Generative Al Engineer is ready to deploy an LLM application written using Foundation Model APIs. They want to follow security best practices for production scenarios

Which authentication method should they choose?

Options:

A.

Use an access token belonging to service principals

B.

Use a frequently rotated access token belonging to either a workspace user or a service principal

C.

Use OAuth machine-to-machine authentication

D.

Use an access token belonging to any workspace user

Buy Now
Questions 10

A company has a typical RAG-enabled, customer-facing chatbot on its website.

Databricks-Generative-AI-Engineer-Associate Question 10

Select the correct sequence of components a user's questions will go through before the final output is returned. Use the diagram above for reference.

Options:

A.

1.embedding model, 2.vector search, 3.context-augmented prompt, 4.response-generating LLM

B.

1.context-augmented prompt, 2.vector search, 3.embedding model, 4.response-generating LLM

C.

1.response-generating LLM, 2.vector search, 3.context-augmented prompt, 4.embedding model

D.

1.response-generating LLM, 2.context-augmented prompt, 3.vector search, 4.embedding model

Buy Now
Questions 11

A Generative Al Engineer would like an LLM to generate formatted JSON from emails. This will require parsing and extracting the following information: order ID, date, and sender email. Here’s a sample email:

Databricks-Generative-AI-Engineer-Associate Question 11

They will need to write a prompt that will extract the relevant information in JSON format with the highest level of output accuracy.

Which prompt will do that?

Options:

A.

You will receive customer emails and need to extract date, sender email, and order ID. You should return the date, sender email, and order ID information in JSON format.

B.

You will receive customer emails and need to extract date, sender email, and order ID. Return the extracted information in JSON format.

Here’s an example: {“date”: “April 16, 2024”, “sender_email”: “sarah.lee925@gmail.com”, “order_id”: “RE987D”}

C.

You will receive customer emails and need to extract date, sender email, and order ID. Return the extracted information in a human-readable format.

D.

You will receive customer emails and need to extract date, sender email, and order ID. Return the extracted information in JSON format.

Buy Now
Questions 12

A Generative Al Engineer has developed an LLM application to answer questions about internal company policies. The Generative AI Engineer must ensure that the application doesn’t hallucinate or leak confidential data.

Which approach should NOT be used to mitigate hallucination or confidential data leakage?

Options:

A.

Add guardrails to filter outputs from the LLM before it is shown to the user

B.

Fine-tune the model on your data, hoping it will learn what is appropriate and not

C.

Limit the data available based on the user’s access level

D.

Use a strong system prompt to ensure the model aligns with your needs.

Buy Now
Questions 13

An AI developer team wants to fine-tune an open-weight model to have exceptional performance on a code generation use case. They are trying to choose the best model to start with. They want to minimize model hosting costs and are using Hugging Face model cards and spaces to explore models. Which TWO model attributes and metrics should the team focus on to make their selection?

Options:

A.

Big Code Models Leaderboard

B.

Number of model parameters

C.

MTEB Leaderboard

D.

Chatbot Arena Leaderboard

E.

Number of model downloads last month

Buy Now
Questions 14

A Generative Al Engineer is developing a RAG application and would like to experiment with different embedding models to improve the application performance.

Which strategy for picking an embedding model should they choose?

Options:

A.

Pick an embedding model trained on related domain knowledge

B.

Pick the most recent and most performant open LLM released at the time

C.

pick the embedding model ranked highest on the Massive Text Embedding Benchmark (MTEB) leaderboard hosted by HuggingFace

D.

Pick an embedding model with multilingual support to support potential multilingual user questions

Buy Now
Questions 15

A Generative AI Engineer at an automotive company would like to build a question-answering chatbot to help customers answer specific questions about their vehicles. They have:

    A catalog with hundreds of thousands of cars manufactured since the 1960s

    Historical searches with user queries and successful matches

    Descriptions of their own cars in multiple languages

They have already selected an open-source LLM and created a test set of user queries. They need to discard techniques that will not help them build the chatbot. Which do they discard?

Options:

A.

Setting chunk size to match the model's context window to maximize coverage

B.

Implementing metadata filtering based on car models and years

C.

Fine-tuning an embedding model on automotive terminology

D.

Adding few-shot examples for response generation

Buy Now
Questions 16

A Generative Al Engineer has already trained an LLM on Databricks and it is now ready to be deployed.

Which of the following steps correctly outlines the easiest process for deploying a model on Databricks?

Options:

A.

Log the model as a pickle object, upload the object to Unity Catalog Volume, register it to Unity Catalog using MLflow, and start a serving endpoint

B.

Log the model using MLflow during training, directly register the model to Unity Catalog using the MLflow API, and start a serving endpoint

C.

Save the model along with its dependencies in a local directory, build the Docker image, and run the Docker container

D.

Wrap the LLM’s prediction function into a Flask application and serve using Gunicorn

Buy Now
Questions 17

A Generative AI Engineer is building a RAG application that will rely on context retrieved from source documents that are currently in PDF format. These PDFs can contain both text and images. They want to develop a solution using the least amount of lines of code.

Which Python package should be used to extract the text from the source documents?

Options:

A.

flask

B.

beautifulsoup

C.

unstructured

D.

numpy

Buy Now
Questions 18

When developing an LLM application, it’s crucial to ensure that the data used for training the model complies with licensing requirements to avoid legal risks.

Which action is NOT appropriate to avoid legal risks?

Options:

A.

Reach out to the data curators directly before you have started using the trained model to let them know.

B.

Use any available data you personally created which is completely original and you can decide what license to use.

C.

Only use data explicitly labeled with an open license and ensure the license terms are followed.

D.

Reach out to the data curators directly after you have started using the trained model to let them know.

Buy Now
Questions 19

Which TWO chain components are required for building a basic LLM-enabled chat application that includes conversational capabilities, knowledge retrieval, and contextual memory?

Options:

A.

(Q)

B.

Vector Stores

C.

Conversation Buffer Memory

D.

External tools

E.

Chat loaders

F.

React Components

Buy Now
Questions 20

A Generative AI Engineer is developing a chatbot designed to assist users with insurance-related queries. The chatbot is built on a large language model (LLM) and is conversational. However, to maintain the chatbot’s focus and to comply with company policy, it must not provide responses to questions about politics. Instead, when presented with political inquiries, the chatbot should respond with a standard message:

“Sorry, I cannot answer that. I am a chatbot that can only answer questions around insurance.”

Which framework type should be implemented to solve this?

Options:

A.

Safety Guardrail

B.

Security Guardrail

C.

Contextual Guardrail

D.

Compliance Guardrail

Buy Now
Questions 21

A Generative Al Engineer is building a production-ready LLM system which replies directly to customers. The solution makes use of the Foundation Model API via provisioned throughput. They are concerned that the LLM could potentially respond in a toxic or otherwise unsafe way. They also wish to perform this with the least amount of effort.

Which approach will do this?

Options:

A.

Host Llama Guard on Foundation Model API and use it to detect unsafe responses

B.

Add some LLM calls to their chain to detect unsafe content before returning text

C.

Add a regex expression on inputs and outputs to detect unsafe responses.

D.

Ask users to report unsafe responses

Buy Now
Exam Name: Databricks Certified Generative AI Engineer Associate
Last Update: Apr 30, 2026
Questions: 73

PDF + Testing Engine

$63.52  $181.49

Testing Engine

$50.57  $144.49
buy now Databricks-Generative-AI-Engineer-Associate testing engine

PDF (Q&A)

$43.57  $124.49
buy now Databricks-Generative-AI-Engineer-Associate pdf