ExamGecko
Question list
Search
Search

List of questions

Search

Related questions











Question 279 - Professional Machine Learning Engineer discussion

Report
Export

You work as an ML researcher at an investment bank and are experimenting with the Gemini large language model (LLM). You plan to deploy the model for an internal use case and need full control of the model's underlying infrastructure while minimizing inference time. Which serving configuration should you use for this task?

A.

Deploy the model on a Vertex AI endpoint using one-click deployment in Model Garden.

Answers
A.

Deploy the model on a Vertex AI endpoint using one-click deployment in Model Garden.

B.

Deploy the model on a Google Kubernetes Engine (GKE) cluster manually by creating a custom YAML manifest.

Answers
B.

Deploy the model on a Google Kubernetes Engine (GKE) cluster manually by creating a custom YAML manifest.

C.

Deploy the model on a Vertex AI endpoint manually by creating a custom inference container.

Answers
C.

Deploy the model on a Vertex AI endpoint manually by creating a custom inference container.

D.

Deploy the model on a Google Kubernetes Engine (GKE) cluster using the deployment options in Model Garden.

Answers
D.

Deploy the model on a Google Kubernetes Engine (GKE) cluster using the deployment options in Model Garden.

Suggested answer: B

Explanation:

Deploying the model on GKE with a custom YAML manifest allows maximum control over infrastructure and latency, aligning with the need for low inference time and internal model use. Vertex AI's one-click deployment (Option A) limits control, and deploying on Vertex AI (Option C) doesn't allow for as much customization as a GKE setup.

asked 07/11/2024
Steve Jenkins-Thorkildson
39 questions
User
Your answer:
0 comments
Sorted by

Leave a comment first