Discount Offer
Available in 1, 3, 6 and 12 Months Free Updates Plans
PDF: $15 $60

Test Engine: $20 $80

PDF + Engine: $25 $99



Pass exam with Dumps4free or we will provide you with three additional months of access for FREE.

Databricks-Generative-AI-Engineer-Associate Practice Test

Whether you're a beginner or brushing up on skills, our Databricks-Generative-AI-Engineer-Associate practice exam is your key to success. Our comprehensive question bank covers all key topics, ensuring you’re fully prepared.


Page 1 out of 6 Pages

A Generative AI Engineer is tasked with deploying an application that takes advantage of a custom MLflow Pyfunc model to return some interim results. How should they configure the endpoint to pass the secrets and credentials?


A. Use spark.conf.set ()


B. Pass variables using the Databricks Feature Store API


C. Add credentials using environment variables


D. Pass the secrets in plain text





C.
  Add credentials using environment variables


Explanation:

Context: Deploying an application that uses an MLflow Pyfunc model involves managing sensitive information such as secrets and credentials securely.
Explanation of Options:

Option A: Use spark.conf.set(): While this method can pass configurations within Spark jobs, using it for secrets is not recommended because it may expose them in logs or Spark UI.

Option B: Pass variables using the Databricks Feature Store API: The Feature Store API is designed for managing features for machine learning, not for handling secrets or credentials.

Option C: Add credentials using environment variables: This is a common practice for managing credentials in a secure manner, as environment variables can be accessed securely by applications without exposing them in the codebase.

Option D: Pass the secrets in plain text: This is highly insecure and not recommended, as it exposes sensitive information directly in the code.

Therefore,Option Cis the best method for securely passing secrets and credentials to an application, protecting them from exposure.

A Generative Al Engineer has developed an LLM application to answer questions about internal company policies. The Generative AI Engineer must ensure that the application doesn’t hallucinate or leak confidential data. Which approach should NOT be used to mitigate hallucination or confidential data leakage?


A. Add guardrails to filter outputs from the LLM before it is shown to the user


B. Fine-tune the model on your data, hoping it will learn what is appropriate and not


C. Limit the data available based on the user’s access level


D. Use a strong system prompt to ensure the model aligns with your needs.





B.
  Fine-tune the model on your data, hoping it will learn what is appropriate and not


Explanation:

When addressing concerns of hallucination and data leakage in an LLM application for internal company policies, fine-tuning the model on internal data with the hope it learns data boundaries can be problematic:

Risk of Data Leakage: Fine-tuning on sensitive or confidential data does not guarantee that the model will not inadvertently include or reference this data in its outputs. There’s a risk of overfitting to the specific data details, which might lead to unintended leakage.

Hallucination: Fine-tuning does not necessarily mitigate the model's tendency to hallucinate; in fact, it might exacerbate it if the training data is not comprehensive or representative of all potential queries.

Better Approaches:

A,C, andDinvolve setting up operational safeguards and constraints that directly address data leakage and ensure responses are aligned with specific user needs and security levels.

Fine-tuning lacks the targeted control needed for such sensitive applications and can introduce new risks, making it an unsuitable approach in this context.

A Generative Al Engineer is tasked with improving the RAG quality by addressing its inflammatory outputs. Which action would be most effective in mitigating the problem of offensive text outputs?


A. Increase the frequency of upstream data updates


B. Inform the user of the expected RAG behavior


C. Restrict access to the data sources to a limited number of users


D. Curate upstream data properly that includes manual review before it is fed into the RAG system





D.
  Curate upstream data properly that includes manual review before it is fed into the RAG system

Explanation:

Addressing offensive or inflammatory outputs in a Retrieval-Augmented Generation (RAG) system is critical for improving user experience and ensuring ethical AI deployment. Here's whyDis the most effective approach:

Manual data curation: The root cause of offensive outputs often comes from the underlying data used to train the model or populate the retrieval system. By manually curating the upstream data and conducting thorough reviews before the data is fed into the RAG system, the engineer can filter out harmful, offensive, or inappropriate content.

Improving data quality: Curating data ensures the system retrieves and generates responses from a high-quality, well-vetted dataset. This directly impacts the relevance and appropriateness of the outputs from the RAG system, preventing inflammatory content from being included in responses.

Effectiveness: This strategy directly tackles the problem at its source (the data) rather than just mitigating the consequences (such as informing users or restricting access). It ensures that the system consistently provides non-offensive, relevant information.

Other options, such as increasing the frequency of data updates or informing users about behavior expectations, may not directly mitigate the generation of inflammatory outputs.

A Generative AI Engineer just deployed an LLM application at a digital marketing company that assists with answering customer service inquiries. Which metric should they monitor for their customer service LLM application in production?


A. Number of customer inquiries processed per unit of time


B. Energy usage per query


C. Final perplexity scores for the training of the model


D. HuggingFace Leaderboard values for the base LLM





A.
  Number of customer inquiries processed per unit of time


Explanation:

When deploying an LLM application for customer service inquiries, the primary focus is on measuring the operational efficiency and quality of the responses. Here's whyAis the correct metric:

Number of customer inquiries processed per unit of time: This metric tracks the throughput of the customer service system, reflecting how many customer inquiries the LLM application can handle in a given time period (e.g., per minute or hour). High throughput is crucial in customer service applications where quick response times are essential to user satisfaction and business efficiency.

Real-time performance monitoring: Monitoring the number of queries processed is an important part of ensuring that the model is performing well under load, especially during peak traffic times. It also helps ensure the system scales properly to meet demand.

Why other options are not ideal:

B. Energy usage per query: While energy efficiency is a consideration, it is not the primary concern for a customer-facing application where user experience (i.e., fast and accurate responses) is critical.

C. Final perplexity scores for the training of the model: Perplexity is a metric for model training, but it doesn't reflect the real-time operational performance of an LLM in production.

D. HuggingFace Leaderboard values for the base LLM: The HuggingFace Leaderboard is more relevant during model selection and benchmarking. However, it is not a direct measure of the model's performance in a specific customer service application in production.

Focusing on throughput (inquiries processed per unit time) ensures that the LLM application is meeting business needs for fast and efficient customer service responses.

A Generative AI Engineer is developing an LLM application that users can use to generate personalized birthday poems based on their names. Which technique would be most effective in safeguarding the application, given the potential for malicious user inputs?


A. Implement a safety filter that detects any harmful inputs and ask the LLM to respond that it is unable to assist


B. Reduce the time that the users can interact with the LLM


C. Ask the LLM to remind the user that the input is malicious but continue the conversation with the user


D. Increase the amount of compute that powers the LLM to process input faster





A.
  Implement a safety filter that detects any harmful inputs and ask the LLM to respond that it is unable to assist


Explanation:

In this case, the Generative AI Engineer is developing an application to generate personalized birthday poems, but there’s a need to safeguard againstmalicious user inputs. The best solution is to implement asafety filter(option A) to detect harmful or inappropriate inputs.

Safety Filter Implementation:Safety filters are essential for screening user input and preventing inappropriate content from being processed by the LLM. These filters can scan inputs for harmful language, offensive terms, or malicious content and intervene before the prompt is passed to the LLM.

Graceful Handling of Harmful Inputs:Once the safety filter detects harmful content, the system can provide a message to the user, such as "I'm unable to assist with this request," instead of processing or responding to malicious input. This protects the system from generating harmful content and ensures a controlled interaction environment.

Why Other Options Are Less Suitable:

B (Reduce Interaction Time): Reducing the interaction time won’t prevent malicious inputs from being entered.

C (Continue the Conversation): While it’s possible to acknowledge malicious input, it is not safe to continue the conversation with harmful content. This could lead to legal or reputational risks.

D (Increase Compute Power): Adding more compute doesn’t address the issue of harmful content and would only speed up processing without resolving safety concerns.

Therefore, implementing asafety filterthat blocks harmful inputs is the most effective technique for safeguarding the application.


Page 1 out of 6 Pages