ExamGecko
Question list
Search
Search

List of questions

Search

Related questions











Question 83 - Certified MuleSoft Platform Architect I discussion

Report
Export

A retail company with thousands of stores has an API to receive data about purchases and insert it into a single database. Each individual store sends a batch of purchase data to the API about every 30 minutes. The API implementation uses a database bulk insert command to submit all the purchase data to a database using a custom JDBC driver provided by a data analytics solution provider. The API implementation is deployed to a single CloudHub worker. The JDBC driver processes the data into a set of several temporary disk files on the CloudHub worker, and then the data is sent to an analytics engine using a proprietary protocol. This process usually takes less than a few minutes. Sometimes a request fails. In this case, the logs show a message from the JDBC driver indicating an out-of-file-space message. When the request is resubmitted, it is successful. What is the best way to try to resolve this throughput issue?

A.
se a CloudHub autoscaling policy to add CloudHub workers
Answers
A.
se a CloudHub autoscaling policy to add CloudHub workers
B.
Use a CloudHub autoscaling policy to increase the size of the CloudHub worker
Answers
B.
Use a CloudHub autoscaling policy to increase the size of the CloudHub worker
C.
Increase the size of the CloudHub worker(s)
Answers
C.
Increase the size of the CloudHub worker(s)
D.
Increase the number of CloudHub workers
Answers
D.
Increase the number of CloudHub workers
Suggested answer: D

Explanation:

Increase the size of the CloudHub worker(s) The key details that we can take out from the given scenario are:>> API implementation uses a database bulk insert command to submit all the purchase data to a database>> JDBC driver processes the data into a set of several temporary disk files on the CloudHub worker>> Sometimes a request fails and the logs show a message indicating an out-of-file-space messageBased on above details:>> Both auto-scaling options does NOT help because we cannot set auto-scaling rules based on error messages. Auto-scaling rules are kicked-off based on CPU/Memory usages and not due to some given error or disk space issues.>> Increasing the number of CloudHub workers also does NOT help here because the reason for the failure is not due to performance aspects w.r.t CPU or Memory. It is due to disk-space.>> Moreover, the API is doing bulk insert to submit the received batch data. Which means, all data is handled by ONE worker only at a time. So, the disk space issue should be tackled on 'per worker' basis. Having multiple workers does not help as the batch may still fail on any worker when disk is out of space on that particular worker.Therefore, the right way to deal this issue and resolve this is to increase the vCore size of the worker so that a new worker with more disk space will be provisioned.

asked 23/09/2024
Aparecido Primo
41 questions
User
Your answer:
0 comments
Sorted by

Leave a comment first