Summer Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: exc65

As an enterprise architect, what are the two reasons for which you would use a canonical data model in the new integration project using Mulesoft Anypoint platform ( choose two answers )

A.

To have consistent data structure aligned in processes

B.

To isolate areas within a bounded context

C.

To incorporate industry standard data formats

D.

There are multiple canonical definitions of each data type

E.

Because the model isolates the back and systems and support mule applications from change

A Mule application uses the Database connector.

What condition can the Mule application automatically adjust to or recover from without needing to restart or redeploy the Mule application?

A.

One of the stored procedures being called by the Mule application has been renamed

B.

The database server was unavailable for four hours due to a major outage but is now fully operational again

C.

The credentials for accessing the database have been updated and the previous credentials are no longer valid

D.

The database server has been updated and hence the database driver library/JAR needs a minor version upgrade

What is not true about Mule Domain Project?

A.

This allows Mule applications to share resources

B.

Expose multiple services within the Mule domain on the same port

C.

Only available Anypoint Runtime Fabric

D.

Send events (messages) to other Mule applications using VM queues

A high-volume eCommerce retailer receives thousands of orders per hour and requires notification of its order management, warehouse, and billing system for subsequent processing within 15 minutes of order submission through its website.

Which integration technology, when used for its typical and intended purpose, meets the retailer’s requirements for this use case?

A.

Managed File Transfer (MFT)

B.

Publish/Subscriber Messaging Bus (Pub/Sub)

C.

Enterprise Data Warehouse (EDW)

D.

Extract Transform Load (ETL)

An organization is designing an integration Mule application to process orders by submitting them to a back-end system for offline processing. Each order will be received by the Mule application through an HTTPS POST and must be acknowledged immediately. Once acknowledged, the order will be submitted to a back-end system. Orders that cannot be successfully submitted due to rejections from the back-end system will need to be processed manually (outside the back-end system).

The Mule application will be deployed to a customer-hosted runtime and is able to use an existing ActiveMQ broker if needed. The ActiveMQ broker is located inside the organization’s firewall. The back-end system has a track record of unreliability due to both minor network connectivity issues and longer outages.

What idiomatic (used for their intended purposes) combination of Mule application components and ActiveMQ queues are required to ensure automatic submission of orders to the back-end system while supporting but minimizing manual order processing?

A.

An Until Successful scope to call the back-end system

One or more ActiveMQ long-retry queues

One or more ActiveMQ dead-letter queues for manual processing

B.

One or more On Error scopes to assist calling the back-end system

An Until Successful scope containing VM components for long retries

A persistent dead-letter VM queue configured in CloudHub

C.

One or more On Error scopes to assist calling the back-end system

One or more ActiveMQ long-retry queues

A persistent dead-letter object store configured in the CloudHub Object Store service

D.

A Batch Job scope to call the back-end system

An Until Successful scope containing Object Store components for long retries

A dead-letter object store configured in the Mule application

An organization has decided on a cloudhub migration strategy that aims to minimize the organizations own IT resources. Currently, the organizational has all of its Mule applications running on its own premises and uses an premises load balancer that exposes all APIs under the base URL https://api.acme.com

As part of the migration strategy, the organization plans to migrate all of its Mule applications and load balancer to cloudhub

What is the most straight-forward and cost effective approach to the Mule applications deployment and load balancing that preserves the public URLs?

A.

Deploy the Mule applications to Cloudhub

Update the CNAME record for an api.acme.com in the organizations DNS server pointing to the A record of a cloudhub dedicated load balancer(DLB)

Apply mapping rules in the DLB to map URLs to their corresponding Mule applications

B.

For each migrated Mule application, deploy an API proxy Mule application to Cloudhub with all applications under the control of a dedicated load balancer(CLB)

Update the CNAME record for api.acme.com in the organization DNS server pointing to the A record of a cloudhub dedicated load balancer(DLB)

Apply mapping rules in the DLB to map each API proxy application to its corresponding Mule applications

C.

Deploy the Mule applications to Cloudhub

Create CNAME record for api.acme.com in the Cloudhub Shared load balancer (SLB) pointing to the A record of the on-premise load balancer

Apply mapping rules in the SLB to map URLs to their corresponding Mule applications

D.

Deploy the Mule applications to Cloudhub

Update the CNAME record for api.acme.com in the organization DNS server pointing to the A record of the cloudhub shared load balancer(SLB)

Apply mapping rules in the SLB to map URLs to their corresponding Mule applications.

A DevOps team has adequate observability of individual system behavior and performance, but it struggles to track the entire lifecycle of each request across different microservices.

Which additional observability approach should this team consider adopting?

A.

Analytics

B.

Metrics

C.

Tracing

D.

Data mining

An organization is building a test suite for their applications using m-unit. The integration architect has recommended using test recorder in studio to record the processing flows and then configure unit tests based on the capture events

What are the two considerations that must be kept in mind while using test recorder

(Choose two answers)

A.

Tests for flows cannot be created with Mule errors raised inside

the flow or already existing in the incoming event

B.

Recorder supports smoking a message before or inside a ForEach processor

C.

The recorder support loops where the structure of the data been tested

changes inside the iteration

D.

A recorded flow execution ends successfully but the result does

not reach its destination because the application is killed

E.

Mocking values resulting from parallel processes are possible and will not affect the execution of the processes that follow in the test

Refer to the exhibit.

A business process involves two APIs that interact with each other asynchronously over HTTP. Each API is implemented as a Mule application. API 1 receives the initial HTTP request and invokes API 2 (in a fire and forget fashion) while API 2, upon completion of the processing, calls back into API l to notify about completion of the asynchronous process.

Each API Is deployed to multiple redundant Mule runtimes and a separate load balancer, and is deployed to a separate network zone.

In the network architecture, how must the firewall rules be configured to enable the above Interaction between API 1 and API 2?

A.

To authorize the certificate to be used both APIs

B.

To enable communication from each API’s Mule Runtimes and Network zone to the load balancer of the other API

C.

To open direct two-way communication between the Mule Runtimes of both API’s

D.

To allow communication between load balancers used by each API

Refer to the exhibit.

A Mule application is deployed to a multi-node Mule runtime cluster. The Mule application uses the competing consumer pattern among its cluster replicas to receive JMS messages from a JMS queue. To process each received JMS message, the following steps are performed in a flow:

Step l: The JMS Correlation ID header is read from the received JMS message.

Step 2: The Mule application invokes an idempotent SOAP webservice over HTTPS, passing the JMS Correlation ID as one parameter in the SOAP request.

Step 3: The response from the SOAP webservice also returns the same JMS Correlation ID.

Step 4: The JMS Correlation ID received from the SOAP webservice is validated to be identical to the JMS Correlation ID received in Step 1.

Step 5: The Mule application creates a response JMS message, setting the JMS Correlation ID message header to the validated JMS Correlation ID and publishes that message to a response JMS queue.

Where should the Mule application store the JMS Correlation ID values received in Step 1 and Step 3 so that the validation in Step 4 can be performed, while also making the overall Mule application highly available, fault-tolerant, performant, and maintainable?

A.

Both Correlation ID values should be stored in a persistent object store

B.

Both Correlation ID values should be stored In a non-persistent object store

C.

The Correlation ID value in Step 1 should be stored in a persistent object store

The Correlation ID value in step 3 should be stored as a Mule event variable/attribute

D.

Both Correlation ID values should be stored as Mule event variable/attribute