Spring Sale 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: Board70

Databricks-Generative-AI-Engineer-Associate Exam Dumps - Databricks Generative AI Engineer Questions and Answers

Question # 14

A Generative Al Engineer would like an LLM to generate formatted JSON from emails. This will require parsing and extracting the following information: order ID, date, and sender email. Here’s a sample email:

They will need to write a prompt that will extract the relevant information in JSON format with the highest level of output accuracy.

Which prompt will do that?

Options:

A.

You will receive customer emails and need to extract date, sender email, and order ID. You should return the date, sender email, and order ID information in JSON format.

B.

You will receive customer emails and need to extract date, sender email, and order ID. Return the extracted information in JSON format.

Here’s an example: {“date”: “April 16, 2024”, “sender_email”: “sarah.lee925@gmail.com”, “order_id”: “RE987D”}

C.

You will receive customer emails and need to extract date, sender email, and order ID. Return the extracted information in a human-readable format.

D.

You will receive customer emails and need to extract date, sender email, and order ID. Return the extracted information in JSON format.

Buy Now
Question # 15

A Generative AI Engineer is testing a simple prompt template in LangChain using the code below, but is getting an error.

Assuming the API key was properly defined, what change does the Generative AI Engineer need to make to fix their chain?

A)

B)

C)

D)

Options:

A.

Option A

B.

Option B

C.

Option C

D.

Option D

Buy Now
Question # 16

A Generative Al Engineer is working with a retail company that wants to enhance its customer experience by automatically handling common customer inquiries. They are working on an LLM-powered Al solution that should improve response times while maintaining a personalized interaction. They want to define the appropriate input and LLM task to do this.

Which input/output pair will do this?

Options:

A.

Input: Customer reviews; Output Group the reviews by users and aggregate per-user average rating, then respond

B.

Input: Customer service chat logs; Output Group the chat logs by users, followed by summarizing each user's interactions, then respond

C.

Input: Customer service chat logs; Output: Find the answers to similar questions and respond with a summary

D.

Input: Customer reviews: Output Classify review sentiment

Buy Now
Question # 17

A Generative AI Engineer is designing a RAG application for answering user questions on technical regulations as they learn a new sport.

What are the steps needed to build this RAG application and deploy it?

Options:

A.

Ingest documents from a source –> Index the documents and saves to Vector Search –> User submits queries against an LLM –> LLM retrieves relevant documents –> Evaluate model –> LLM generates a response –> Deploy it using Model Serving

B.

Ingest documents from a source –> Index the documents and save to Vector Search –> User submits queries against an LLM –> LLM retrieves relevant documents –> LLM generates a response -> Evaluate model –> Deploy it using Model Serving

C.

Ingest documents from a source –> Index the documents and save to Vector Search –> Evaluate model –> Deploy it using Model Serving

D.

User submits queries against an LLM –> Ingest documents from a source –> Index the documents and save to Vector Search –> LLM retrieves relevant documents –> LLM generates a response –> Evaluate model –> Deploy it using Model Serving

Buy Now
Question # 18

A team wants to serve a code generation model as an assistant for their software developers. It should support multiple programming languages. Quality is the primary objective.

Which of the Databricks Foundation Model APIs, or models available in the Marketplace, would be the best fit?

Options:

A.

Llama2-70b

B.

BGE-large

C.

MPT-7b

D.

CodeLlama-34B

Buy Now
Question # 19

Generative AI Engineer at an electronics company just deployed a RAG application for customers to ask questions about products that the company carries. However, they received feedback that the RAG response often returns information about an irrelevant product.

What can the engineer do to improve the relevance of the RAG’s response?

Options:

A.

Assess the quality of the retrieved context

B.

Implement caching for frequently asked questions

C.

Use a different LLM to improve the generated response

D.

Use a different semantic similarity search algorithm

Buy Now
Question # 20

A Generative Al Engineer is using an LLM to classify species of edible mushrooms based on text descriptions of certain features. The model is returning accurate responses in testing and the Generative Al Engineer is confident they have the correct list of possible labels, but the output frequently contains additional reasoning in the answer when the Generative Al Engineer only wants to return the label with no additional text.

Which action should they take to elicit the desired behavior from this LLM?

Options:

A.

Use few snot prompting to instruct the model on expected output format

B.

Use zero shot prompting to instruct the model on expected output format

C.

Use zero shot chain-of-thought prompting to prevent a verbose output format

D.

Use a system prompt to instruct the model to be succinct in its answer

Buy Now
Question # 21

All of the following are Python APIs used to query Databricks foundation models. When running in an interactive notebook, which of the following libraries does not automatically use the current session credentials?

Options:

A.

OpenAI client

B.

REST API via requests library

C.

MLflow Deployments SDK

D.

Databricks Python SDK

Buy Now
Question # 22

What is an effective method to preprocess prompts using custom code before sending them to an LLM?

Options:

A.

Directly modify the LLM’s internal architecture to include preprocessing steps

B.

It is better not to introduce custom code to preprocess prompts as the LLM has not been trained with examples of the preprocessed prompts

C.

Rather than preprocessing prompts, it’s more effective to postprocess the LLM outputs to align the outputs to desired outcomes

D.

Write a MLflow PyFunc model that has a separate function to process the prompts

Buy Now
Question # 23

A Generative AI Engineer is tasked with deploying an application that takes advantage of a custom MLflow Pyfunc model to return some interim results.

How should they configure the endpoint to pass the secrets and credentials?

Options:

A.

Use spark.conf.set ()

B.

Pass variables using the Databricks Feature Store API

C.

Add credentials using environment variables

D.

Pass the secrets in plain text

Buy Now
Exam Name: Databricks Certified Generative AI Engineer Associate
Last Update: Mar 15, 2026
Questions: 73
Databricks-Generative-AI-Engineer-Associate pdf

Databricks-Generative-AI-Engineer-Associate PDF

$25.5  $84.99
Databricks-Generative-AI-Engineer-Associate Engine

Databricks-Generative-AI-Engineer-Associate Testing Engine

$28.5  $94.99
Databricks-Generative-AI-Engineer-Associate PDF + Engine

Databricks-Generative-AI-Engineer-Associate PDF + Testing Engine

$40.5  $134.99