ExamGecko
Question list
Search
Search

List of questions

Search

Related questions











Question 285 - Professional Machine Learning Engineer discussion

Report
Export

Your organization's marketing team is building a customer recommendation chatbot that uses a generative AI large language model (LLM) to provide personalized product suggestions in real time. The chatbot needs to access data from millions of customers, including purchase history, browsing behavior, and preferences. The data is stored in a Cloud SQL for PostgreSQL database. You need the chatbot response time to be less than 100ms. How should you design the system?

A.

Use BigQuery ML to fine-tune the LLM with the data in the Cloud SQL for PostgreSQL database, and access the model from BigQuery.

Answers
A.

Use BigQuery ML to fine-tune the LLM with the data in the Cloud SQL for PostgreSQL database, and access the model from BigQuery.

B.

Replicate the Cloud SQL for PostgreSQL database to AlloyDB. Configure the chatbot server to query AlloyDB.

Answers
B.

Replicate the Cloud SQL for PostgreSQL database to AlloyDB. Configure the chatbot server to query AlloyDB.

C.

Transform relevant customer data into vector embeddings and store them in Vertex AI Search for retrieval by the LLM.

Answers
C.

Transform relevant customer data into vector embeddings and store them in Vertex AI Search for retrieval by the LLM.

D.

Create a caching layer between the chatbot and the Cloud SQL for PostgreSQL database to store frequently accessed customer data. Configure the chatbot server to query the cache.

Answers
D.

Create a caching layer between the chatbot and the Cloud SQL for PostgreSQL database to store frequently accessed customer data. Configure the chatbot server to query the cache.

Suggested answer: D

Explanation:

A caching layer is essential to reduce database access time, meeting the <100ms requirement. Caches store high-frequency, low-latency queries in memory, minimizing access delays caused by database lookups. While AlloyDB (Option B) provides performance benefits, a caching layer is more efficient and cost-effective for this purpose. BigQuery ML (Option A) is less ideal for real-time personalized responses due to access speed, and vector embeddings (Option C) are not needed unless semantic search is a requirement.

asked 07/11/2024
ayodele fakayode
35 questions
User
Your answer:
0 comments
Sorted by

Leave a comment first