ExamGecko
Home Home / Salesforce / Certified MuleSoft Integration Architect I

Salesforce Certified MuleSoft Integration Architect I Practice Test - Questions Answers, Page 9

Question list
Search
Search

List of questions

Search

Related questions











An organization is designing a mule application to support an all or nothing transaction between serval database operations and some other connectors so that they all roll back if there is a problem with any of the connectors

Besides the database connector , what other connector can be used in the transaction.

A.
VM
A.
VM
Answers
B.
Anypoint MQ
B.
Anypoint MQ
Answers
C.
SFTP
C.
SFTP
Answers
D.
ObjectStore
D.
ObjectStore
Answers
Suggested answer: A

Explanation:

Correct answer is VM VM support Transactional Type. When an exception occur, The transaction rolls back to its original state for reprocessing. This feature is not supported by other connectors.

Here is additional information about Transaction management:

A mule application uses an HTTP request operation to involve an external API.

The external API follows the HTTP specification for proper status code usage.

What is possible cause when a 3xx status code is returned to the HTTP Request operation from the external API?

A.
The request was not accepted by the external API
A.
The request was not accepted by the external API
Answers
B.
The request was Redirected to a different URL by the external API
B.
The request was Redirected to a different URL by the external API
Answers
C.
The request was NOT RECEIVED by the external API
C.
The request was NOT RECEIVED by the external API
Answers
D.
The request was ACCEPTED by the external API
D.
The request was ACCEPTED by the external API
Answers
Suggested answer: B

Explanation:

3xx HTTP status codes indicate a redirection that the user agent (a web browser or a crawler) needs to take further action when trying to access a particular resource.

An organization is migrating all its Mule applications to Runtime Fabric (RTF). None of the Mule applications use Mule domain projects.

Currently, all the Mule applications have been manually deployed to a server group among several customer hosted Mule runtimes.

Port conflicts between these Mule application deployments are currently managed by the DevOps team who carefully manage Mule application properties files.

When the Mule applications are migrated from the current customer-hosted server group to Runtime Fabric (RTF), fo the Mule applications need to be rewritten and what DevOps port configuration responsibilities change or stay the same?

A.
Yes, the Mule applications Must be rewritten DevOps No Longer needs to manage port conflicts between the Mule applications
A.
Yes, the Mule applications Must be rewritten DevOps No Longer needs to manage port conflicts between the Mule applications
Answers
B.
Yes, the Mule applications Must be rewritten DevOps Must Still Manage port conflicts.
B.
Yes, the Mule applications Must be rewritten DevOps Must Still Manage port conflicts.
Answers
C.
NO, The Mule applications do NOT need to be rewritten DevOps MUST STILL manage port conflicts
C.
NO, The Mule applications do NOT need to be rewritten DevOps MUST STILL manage port conflicts
Answers
D.
NO, the Mule applications do NO need to be rewritten DevOps NO LONGER needs to manage port conflicts between the Mule applications.
D.
NO, the Mule applications do NO need to be rewritten DevOps NO LONGER needs to manage port conflicts between the Mule applications.
Answers
Suggested answer: C

Explanation:

* Anypoint Runtime Fabric is a container service that automates the deployment and orchestration of your Mule applications and gateways.

* Runtime Fabric runs on customer-managed infrastructure on AWS, Azure, virtual machines (VMs) or bare-metal servers.

* As none of the Mule applications use Mule domain projects. applications are not required to be rewritten. Also when applications are deployed on RTF, by default ingress is allowed only on 8081.

* Hence port conflicts are not required to be managed by DevOps team

An organization is evaluating using the CloudHub shared Load Balancer (SLB) vs creating a CloudHub dedicated load balancer (DLB). They are evaluating how this choice affects the various types of certificates used by CloudHub deplpoyed Mule applications, including MuleSoft-provided, customer-provided, or Mule application-provided certificates.

What type of restrictions exist on the types of certificates that can be exposed by the CloudHub Shared Load Balancer (SLB) to external web clients over the public internet?

A.
Only MuleSoft-provided certificates are exposed.
A.
Only MuleSoft-provided certificates are exposed.
Answers
B.
Only customer-provided wildcard certificates are exposed.
B.
Only customer-provided wildcard certificates are exposed.
Answers
C.
Only customer-provided self-signed certificates are exposed.
C.
Only customer-provided self-signed certificates are exposed.
Answers
D.
Only underlying Mule application certificates are exposed (pass-through)
D.
Only underlying Mule application certificates are exposed (pass-through)
Answers
Suggested answer: A

Explanation:

https://docs.mulesoft.com/runtime-manager/dedicated-load-balancer-tutorial

A Mule application is being designed To receive nightly a CSV file containing millions of records from an external vendor over SFTP, The records from the file need to be validated, transformed. And then written to a database. Records can be inserted into the database in any order.

In this use case, what combination of Mule components provides the most effective and performant way to write these records to the database?

A.
Use a Parallel for Each scope to Insert records one by one into the database
A.
Use a Parallel for Each scope to Insert records one by one into the database
Answers
B.
Use a Scatter-Gather to bulk insert records into the database
B.
Use a Scatter-Gather to bulk insert records into the database
Answers
C.
Use a Batch job scope to bulk insert records into the database.
C.
Use a Batch job scope to bulk insert records into the database.
Answers
D.
Use a DataWeave map operation and an Async scope to insert records one by one into the database.
D.
Use a DataWeave map operation and an Async scope to insert records one by one into the database.
Answers
Suggested answer: C

Explanation:

Correct answer is Use a Batch job scope to bulk insert records into the database

* Batch Job is most efficient way to manage millions of records.

A few points to note here are as follows :

Reliability: If you want reliabilty while processing the records, i.e should the processing survive a runtime crash or other unhappy scenarios, and when restarted process all the remaining records, if yes then go for batch as it uses persistent queues.

Error Handling: In Parallel for each an error in a particular route will stop processing the remaining records in that route and in such case you'd need to handle it using on error continue, batch process does not stop during such error instead you can have a step for failures and have a dedicated handling in it.

Memory footprint: Since question said that there are millions of records to process, parallel for each will aggregate all the processed records at the end and can possibly cause Out Of Memory.

Batch job instead provides a BatchResult in the on complete phase where you can get the count of failures and success. For huge file processing if order is not a concern definitely go ahead with Batch Job

An automation engineer needs to write scripts to automate the steps of the API lifecycle, including steps to create, publish, deploy and manage APIs and their implementations in Anypoint Platform.

What Anypoint Platform feature can be used to automate the execution of all these actions in scripts in the easiest way without needing to directly invoke the Anypoint Platform REST APIs?

A.
Automated Policies in API Manager
A.
Automated Policies in API Manager
Answers
B.
Runtime Manager agent
B.
Runtime Manager agent
Answers
C.
The Mule Maven Plugin
C.
The Mule Maven Plugin
Answers
D.
Anypoint CLI
D.
Anypoint CLI
Answers
Suggested answer: D

Explanation:

Anypoint Platform provides a scripting and command-line tool for both Anypoint Platform and Anypoint Platform Private Cloud Edition (Anypoint Platform PCE). The command-line interface (CLI) supports both the interactive shell and standard CLI modes and works with: Anypoint Exchange Access management Anypoint Runtime Manager

A company wants its users to log in to Anypoint Platform using the company's own internal user credentials. To achieve this, the company needs to integrate an external identity provider (IdP) with the company's Anypoint Platform master organization, but SAML 2.0 CANNOT be used. Besides SAML 2.0, what single-sign-on standard can the company use to integrate the IdP with their Anypoint Platform master organization?

A.
SAML 1.0
A.
SAML 1.0
Answers
B.
OAuth 2.0
B.
OAuth 2.0
Answers
C.
Basic Authentication
C.
Basic Authentication
Answers
D.
OpenID Connect
D.
OpenID Connect
Answers
Suggested answer: D

Explanation:

As the Anypoint Platform organization administrator, you can configure identity management in Anypoint Platform to set up users for single sign-on (SSO).

Configure identity management using one of the following single sign-on standards:

1) OpenID Connect: End user identity verification by an authorization server including SSO

2) SAML 2.0: Web-based authorization including cross-domain SSO

An API implementation is being developed to expose data from a production database via HTTP requests. The API implementation executes a database SELECT statement that is dynamically created based upon data received from each incoming HTTP request. The developers are planning to use various types of testing to make sure the Mule application works as expected, can handle specific workloads, and behaves correctly from an API consumer perspective. What type of testing would typically mock the results from each SELECT statement rather than actually execute it in the production database?

A.
Unit testing (white box)
A.
Unit testing (white box)
Answers
B.
Integration testing
B.
Integration testing
Answers
C.
Functional testing (black box)
C.
Functional testing (black box)
Answers
D.
Performance testing
D.
Performance testing
Answers
Suggested answer: A

Explanation:

In Unit testing instead of using actual backends, stubs are used for the backend services. This ensures that developers are not blocked and have no dependency on other systems.

In Unit testing instead of using actual backends, stubs are used for the backend services. This ensures that developers are not blocked and have no dependency on other systems.

Below are the typical characteristics of unit testing.

-- Unit tests do not require deployment into any special environment, such as a staging environment

-- Unit tests san be run from within an embedded Mule runtime

-- Unit tests can/should be implemented using MUnit

-- For read-only interactions to any dependencies (such as other APIs): allowed to invoke production endpoints

-- For write interactions: developers must implement mocks using MUnit

-- Require knowledge of the implementation details of the API implementation under test

A travel company wants to publish a well-defined booking service API to be shared with its business partners. These business partners have agreed to ONLY consume SOAP services and they want to get the service contracts in an easily consumable way before they start any development. The travel company will publish the initial design documents to Anypoint Exchange, then share those documents with the business partners. When using an API-led approach, what is the first design document the travel company should deliver to its business partners?

A.
Create a WSDL specification using any XML editor
A.
Create a WSDL specification using any XML editor
Answers
B.
Create a RAML API specification using any text editor
B.
Create a RAML API specification using any text editor
Answers
C.
Create an OAS API specification in Design Center
C.
Create an OAS API specification in Design Center
Answers
D.
Create a SOAP API specification in Design Center
D.
Create a SOAP API specification in Design Center
Answers
Suggested answer: A

Explanation:

SOAPAPI specifications are provided as WSDL. Design center doesn't provide the functionality to create WSDL file. Hence WSDL needs to be created using XML editor

What is not true about Mule Domain Project?

A.
This allows Mule applications to share resources
A.
This allows Mule applications to share resources
Answers
B.
Expose multiple services within the Mule domain on the same port
B.
Expose multiple services within the Mule domain on the same port
Answers
C.
Only available Anypoint Runtime Fabric
C.
Only available Anypoint Runtime Fabric
Answers
D.
Send events (messages) to other Mule applications using VM queues
D.
Send events (messages) to other Mule applications using VM queues
Answers
Suggested answer: C

Explanation:

*Mule Domain Project is ONLY available for customer-hosted Mule runtimes, but not for Anypoint Runtime Fabric

*Mule domain project is available for Hybrid and Private Cloud (PCE). Rest all provide application isolation and can't support domain project.

What is Mule Domain Project?

*A Mule Domain Project is implemented to configure the resources that are shared among different projects. These resources can be used by all the projects associated with this domain. Mule applications can be associated with only one domain, but a domain can be associated with multiple projects. Shared resources allow multiple development teams to work in parallel using the same set of reusable connectors. Defining these connectors as shared resources at the domain level allows the team to: - Expose multiple services within the domain through the same port. - Share the connection to persistent storage. - Share services between apps through a well-defined interface. - Ensure consistency between apps upon any changes because the configuration is only set in one place.

*Use domains Project to share the same host and port among multiple projects. You can declare the http connector within a domain project and associate the domain project with other projects. Doing this also allows to control thread settings, keystore configurations, time outs for all the requests made within multiple applications. You may think that one can also achieve this by duplicating the http connector configuration across all the applications. But, doing this may pose a nightmare if you have to make a change and redeploy all the applications.

*If you use connector configuration in the domain and let all the applications use the new domain instead of a default domain, you will maintain only one copy of the http connector configuration. Any changes will require only the domain to the redeployed instead of all the applications.

You can start using domains in only three steps:

1) Create a Mule Domain project

2) Create the global connector configurations which needs to be shared across the applications inside the Mule Domain project

3) Modify the value of domain in mule-deploy.properties file of the applications

Total 273 questions
Go to page: of 28