Deal of The Day! Hurry Up, Grab the Special Discount - Save 25% - Ends In 00:00:00 Coupon code: SAVE25
Welcome to Pass4Success

- Free Preparation Discussions

Databricks Exam Databricks Certified Generative AI Engineer Associate Topic 4 Question 20 Discussion

Actual exam question for Databricks's Databricks Certified Generative AI Engineer Associate exam
Question #: 20
Topic #: 4
[All Databricks Certified Generative AI Engineer Associate Questions]

A Generative Al Engineer is developing a RAG system for their company to perform internal document Q&A for structured HR policies, but the answers returned are frequently incomplete and unstructured It seems that the retriever is not returning all relevant context The Generative Al Engineer has experimented with different embedding and response generating LLMs but that did not improve results.

Which TWO options could be used to improve the response quality?

Choose 2 answers

Show Suggested Answer Hide Answer
Suggested Answer: A, B

The problem describes a Retrieval-Augmented Generation (RAG) system for HR policy Q&A where responses are incomplete and unstructured due to the retriever failing to return sufficient context. The engineer has already tried different embedding and response-generating LLMs without success, suggesting the issue lies in the retrieval process---specifically, how documents are chunked and indexed. Let's evaluate the options.

Option A: Add the section header as a prefix to chunks

Adding section headers provides additional context to each chunk, helping the retriever understand the chunk's relevance within the document structure (e.g., ''Leave Policy: Annual Leave'' vs. just ''Annual Leave''). This can improve retrieval precision for structured HR policies.

Databricks Reference: 'Metadata, such as section headers, can be appended to chunks to enhance retrieval accuracy in RAG systems' ('Databricks Generative AI Cookbook,' 2023).

Option B: Increase the document chunk size

Larger chunks include more context per retrieval, reducing the chance of missing relevant information split across smaller chunks. For structured HR policies, this can ensure entire sections or rules are retrieved together.

Databricks Reference: 'Increasing chunk size can improve context completeness, though it may trade off with retrieval specificity' ('Building LLM Applications with Databricks').

Option C: Split the document by sentence

Splitting by sentence creates very small chunks, which could exacerbate the problem by fragmenting context further. This is likely why the current system fails---it retrieves incomplete snippets rather than cohesive policy sections.

Databricks Reference: No specific extract opposes this, but the emphasis on context completeness in RAG suggests smaller chunks worsen incomplete responses.

Option D: Use a larger embedding model

A larger embedding model might improve vector quality, but the question states that experimenting with different embedding models didn't help. This suggests the issue isn't embedding quality but rather chunking/retrieval strategy.

Databricks Reference: Embedding models are critical, but not the focus when retrieval context is the bottleneck.

Option E: Fine tune the response generation model

Fine-tuning the LLM could improve response coherence, but if the retriever doesn't provide complete context, the LLM can't generate full answers. The root issue is retrieval, not generation.

Databricks Reference: Fine-tuning is recommended for domain-specific generation, not retrieval fixes ('Generative AI Engineer Guide').

Conclusion: Options A and B address the retrieval issue directly by enhancing chunk context---either through metadata (A) or size (B)---aligning with Databricks' RAG optimization strategies. C would worsen the problem, while D and E don't target the root cause given prior experimentation.


Contribute your Thoughts:

Kindra
2 days ago
I feel like splitting the document by sentence could make things more fragmented, but it might also help with clarity in some cases.
upvoted 0 times
...
Kina
8 days ago
I'm not entirely sure, but I think increasing the document chunk size might lead to more comprehensive responses.
upvoted 0 times
...
Mitsue
13 days ago
I remember discussing the importance of context in our last practice session, so maybe adding the section header as a prefix could help clarify the answers.
upvoted 0 times
...
Stephaine
18 days ago
This seems like a classic case of needing to experiment and see what works best. I'd probably start by trying the section header prefix and fine-tuning the model. Those seem like the most straightforward options to test out first. Gotta love these tricky AI engineering problems!
upvoted 0 times
...
Yuriko
23 days ago
I'm a bit confused on this one. The question mentions the retriever isn't returning all the relevant context, so I'm not sure if splitting by sentence or increasing chunk size is the way to go. Maybe adding the section headers or fine-tuning the model would be a better approach to try first.
upvoted 0 times
...
Ivette
28 days ago
Okay, let's see here. I'm leaning towards using a larger embedding model - that might help capture more of the relevant information in the documents. And splitting the document by sentence could be worth a try too, to get more granular context.
upvoted 0 times
...
Alverta
1 month ago
Hmm, this seems like a tricky one. I'm thinking adding the section headers as a prefix could help provide more context, but I'm not sure if that's the best approach. Maybe increasing the chunk size or fine-tuning the response generation model could also be good options.
upvoted 0 times
...
Amos
1 month ago
Increasing the document chunk size seems like a no-brainer to me. More data to work with, more chances to find the right answers.
upvoted 0 times
...
Elli
2 months ago
I think adding the section header as a prefix to chunks could really help provide more context for the responses. It's like giving the AI a roadmap to follow.
upvoted 0 times
...

Save Cancel