New Year Sale 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: 70track

Free Salesforce MuleSoft-Integration-Architect-I Practice Exam with Questions & Answers | Set: 7

Questions 61

An organization has implemented the cluster with two customer hosted Mule runtimes is hosting an application.

This application has a flow with a JMS listener configured to consume messages from a queue destination. As an integration architect can you advise which JMS listener configuration must be used to receive messages in all the nodes of the cluster?

Options:
A.

Use the parameter primaryNodeOnly= "false" on the JMS listener

B.

Use the parameter primaryNodeOnly= "false" on the JMS listener with a shared subscription

C.

Use the parameter primaryNodeOnly= "true" on the JMS listener with a non­shared subscription

D.

Use the parameter primaryNodeOnly= "true" on the JMS listener

Salesforce MuleSoft-Integration-Architect-I Premium Access
Questions 62

Refer to the exhibit.

A Mule application is deployed to a multi-node Mule runtime cluster. The Mule application uses the competing consumer pattern among its cluster replicas to receive JMS messages from a JMS queue. To process each received JMS message, the following steps are performed in a flow:

Step l: The JMS Correlation ID header is read from the received JMS message.

Step 2: The Mule application invokes an idempotent SOAP webservice over HTTPS, passing the JMS Correlation ID as one parameter in the SOAP request.

Step 3: The response from the SOAP webservice also returns the same JMS Correlation ID.

Step 4: The JMS Correlation ID received from the SOAP webservice is validated to be identical to the JMS Correlation ID received in Step 1.

Step 5: The Mule application creates a response JMS message, setting the JMS Correlation ID message header to the validated JMS Correlation ID and publishes that message to a response JMS queue.

Where should the Mule application store the JMS Correlation ID values received in Step 1 and Step 3 so that the validation in Step 4 can be performed, while also making the overall Mule application highly available, fault-tolerant, performant, and maintainable?

Options:
A.

Both Correlation ID values should be stored in a persistent object store

B.

Both Correlation ID values should be stored In a non-persistent object store

C.

The Correlation ID value in Step 1 should be stored in a persistent object storeThe Correlation ID value in step 3 should be stored as a Mule event variable/attribute

D.

Both Correlation ID values should be stored as Mule event variable/attribute

Questions 63

An external web UI application currently accepts occasional HTTP requests from client web browsers to change (insert, update, or delete) inventory pricing information in an inventory system's database. Each inventory pricing change must be transformed and then synchronized with multiple customer experience systems in near real-time (in under 10 seconds). New customer experience systems are expected to be added in the future.

The database is used heavily and limits the number of SELECT queries that can be made to the database to 10 requests per hour per user.

What is the most scalable, idiomatic (used for its intended purpose), decoupled. reusable, and maintainable integration mechanism available to synchronize each inventory pricing change with the various customer experience systems in near real-time?

Options:
A.

Write a Mule application with a Database On Table Row event source configured for the inventory pricing database, with the watermark attribute set to an appropriate database columnIn the same now, use a Scatter-Gather to call each customer experience system's REST API with transformed inventory-pricing records

B.

Add a trigger to the inventory-pricing database table so that for each change to the inventory pricing database, a stored procedure is called that makes a REST call to a Mule applicationWrite the Mule application to publish each Mule event as a message to an Anypoint MQ exchangeWrite other Mule applications to subscribe to the Anypoint MQ exchange, transform each received message, and then update the Mule application's corresponding custome

C.

Replace the external web UI application with a Mule application to accept HTTP requests from client web browsersIn the same Mule application, use a Batch Job scope to test if the database request will succeed, aggregate pricing changes within a short time window, and then update both the inventory pricing database and each customer experience system using a Parallel For Each scope

D.

Write a Mule application with a Database On Table Row event source configured for the inventory pricing database, with the ID attribute set to an appropriate database columnIn the same flow, use a Batch Job scope to publish transformed Inventory-pricing records to an Anypoint MQ queueWrite other Mule applications to subscribe to the Anypoint MQ queue, transform each received message, and then update the Mule application's corresponding cust

Questions 64

Refer to the exhibit.

MuleSoft-Integration-Architect-I Question 64

One of the backend systems invoked by an API implementation enforces rate limits on the number of requests a particular client can make. Both the backend system and the API implementation are deployed to several non-production environments in addition to production.

Rate limiting of the backend system applies to all non-production environments. The production environment, however, does NOT have any rate limiting.

What is the most effective approach to conduct performance tests of the API implementation in a staging (non-production) environment?

Options:
A.

Create a mocking service that replicates the backend system's production performance characteristics. Then configure the API implementation to use the mocking service and conduct the performance tests

B.

Use MUnit to simulate standard responses from the backend system then conduct performance tests to identify other bottlenecks in the system

C.

Include logic within the API implementation that bypasses invocations of the backend system in a performance test situation. Instead invoking local stubs that replicate typical backend system responses then conduct performance tests using this API Implementation

D.

Conduct scaled-down performance tests in the staging environment against the rate limited backend system then upscale performance results to full production scale

Questions 65

An organization has decided on a cloudhub migration strategy that aims to minimize the organizations own IT resources. Currently, the organizational has all of its Mule applications running on its own premises and uses an premises load balancer that exposes all APIs under the base URL https://api.acme.com

As part of the migration strategy, the organization plans to migrate all of its Mule applications and load balancer to cloudhub

What is the most straight-forward and cost effective approach to the Mule applications deployment and load balancing that preserves the public URLs?

Options:
A.

Deploy the Mule applications to CloudhubUpdate the CNAME record for an api.acme.com in the organizations DNS server pointing to the A record of a cloudhub dedicated load balancer(DLB)Apply mapping rules in the DLB to map URLs to their corresponding Mule applications

B.

For each migrated Mule application, deploy an API proxy Mule application to Cloudhub with all applications under the control of a dedicated load balancer(CLB)Update the CNAME record for api.acme.com in the organization DNS server pointing to the A record of a cloudhub dedicated load balancer(DLB)Apply mapping rules in the DLB to map each API proxy application to its corresponding Mule applications

C.

Deploy the Mule applications to CloudhubCreate CNAME record for api.acme.com in the Cloudhub Shared load balancer (SLB) pointing to the A record of the on-premise load balancerApply mapping rules in the SLB to map URLs to their corresponding Mule applications

D.

Deploy the Mule applications to CloudhubUpdate the CNAME record for api.acme.com in the organization DNS server pointing to the A record of the cloudhub shared load balancer(SLB)Apply mapping rules in the SLB to map URLs to their corresponding Mule applications.

Questions 66

A bank is implementing a REST API in a Mule application to receive an array of accounts from an online banking platform user interface (UI),retrieve account balances for those accounts from a backend Finance system, and then return the account balances so they can be displayed inthe online banking platform UI. As part of the processing, the MuleSoft API also needs to insert the retrieved account data into an AuditDatabase for auditing purposes. the auditing process should not add latency to the account balance retrieval response back to the onlinebanking platform UI.

The retrieveBalances flow in the Mule application is designed to use an operation in a connector to the Finance system (the Finance operation) that

can only look up one account record at a time, and a operation from a different connector to the Audit system (the Audit operation) that can only

insert one account record at a time.

To best meet the performance-related requirements, what scope or scopes should be used and how should they be used to incorporate the Finance

operation and Audit operation into the retrieveBalances flow?

Options:
A.

Wrap the Finance operation in a Parallel For-Each scope. Wrap the Audit operation in a Async scope.

B.

Wrap the Finance operation in a Until-Successful scope. Wrap the Audit operation in a Try-Catch scope.

C.

Wrap both connector operations in a Async scope.

D.

Wrap both connector operations in a For-Each scope.

Questions 67

According to MuleSoft, which deployment characteristic applies to a microservices application architecture?

Options:
A.

Services exist as independent deployment artifacts and can be scaled -independently of other services

B.

All services of an application can be deployed together as single Java WAR file

C.

A deployment to enhance one capability requires a redeployment of all capabilities

D.

Core business capabilities are encapsulated in a single, deployable application

Questions 68

A company is implementing a new Mule application that supports a set of critical functions driven by a rest API enabled, claims payment rules engine hosted on oracle ERP. As designed the mule application requires many data transformation operations as it performs its batch processing logic.

The company wants to leverage and reuse as many of its existing java-based capabilities (classes, objects, data model etc.) as possible

What approach should be considered when implementing required data mappings and transformations between Mule application and Oracle ERP in the new Mule application?

Options:
A.

Create a new metadata RAML classes in Mule from the appropriate Java objects and then perform transformations via Dataweave

B.

From the mule application, transform via theXSLT model

C.

Transform by calling any suitable Java class from Dataweave

D.

Invoke any of the appropriate Java methods directly, create metadata RAML classes and then perform required transformations via Dataweave

Questions 69

An application load balancer routes requests to a RESTful web API secured by Anypoint Flex Gateway.

Which protocol is involved in the communication between the load balancer and the Gateway?

Options:
A.

SFTP

B.

HTTPS

C.

LDAP

D.

SMTP

Questions 70

According to MuleSoft, what Action should an IT organization take regarding its technology assets in order to close the IT delivery.

Options:
A.

Make assets easily discoverable via a central repository

B.

Focus project delivery efforts on custom assets that meet the specific requirements of each individual line of business

C.

Create weekly meetings that all members of IT attend to present justification and request approval to use existing assets

D.

Hire additional staff to meet the demand for asset creation required for approved projects and timelines