Summer Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: exc65

Every year, Ursa Major Solar has more than 1 million orders. Each order contains an average of 10 line items. The Chief Executive Officer (CEO) needs the Sales Reps to see how much money each customer generates year-over-year. However, data storage is running low in Salesforce.

Which approach for data archiving is appropriate for this scenario?

A.

1. Annually export and delete order line items. 2. Store them in a zip file in case the data is needed later.

B.

1. Annually aggregate order amount data to store in a custom object. 2. Delete those orders and order line items.

C.

1. Annually export and delete orders and order line items. 2. Store them in a zip file in case the data is needed later.

D.

1. Annually delete orders and order line items. 2. Ensure the customer has order information in another system.

To avoid creating duplicate Contacts, a customer frequently uses Data Loader to upsert Contact records into Salesforce. What common error should the data architect be aware of when using upsert?

A.

Errors with duplicate external Id values within the same CSV file.

B.

Errors with records being updated and inserted in the same CSV file.

C.

Errors when a duplicate Contact name is found cause upsert to fail.

D.

Errors with using the wrong external Id will cause the load to fail.

What 2 data management policies does the data classification feature allow customers to classify in salesforce? Choose 2 answers:

A.

Reference data policy.

B.

Data governance policy.

C.

Data sensitivity level

D.

Compliance categorization policy.

DreamHouse Realty has 15 million records in the Order_c custom object. When running a bulk query, the query times out.

What should be considered to address this issue?

A.

Tooling API

B.

PK Chunking

C.

Metadata API

D.

Streaming API

Universal Containers (UC) is launching an RFP to acquire a new accounting product available on AppExchange. UC is expecting to issue 5 million invoices per year, with each invoice containing an average of 10 line items. What should UC's Data Architect recommend to ensure scalability?

A.

Ensure invoice line items simply reference existing Opportunity line items.

B.

Ensure the account product vendor includes Wave Analytics in their offering.

C.

Ensure the account product vendor provides a sound data archiving strategy.

D.

Ensure the accounting product runs 100% natively on the Salesforce platform.

Northern Trail Outfitters needs to implement an archive solution for Salesforce data. This archive solution needs to help NTO do the following:

1. Remove outdated Information not required on a day-to-day basis.

2. Improve Salesforce performance.

Which solution should be used to meet these requirements?

A.

Identify a location to store archived data and use scheduled batch jobs to migrate and purge the aged data on a nightly basis,

B.

Identify a location to store archived data, and move data to the location using a time-based workflow.

C.

Use a formula field that shows true when a record reaches a defined age and use that field to run a report and export a report into SharePoint.

D.

Create a full copy sandbox, and use it as a source for retaining archived data.

UC is implementing sales cloud for patient management and would like to encrypt sensitive patient records being stored in files.

Which solution should a data architect recommend to solve this requirement?

A.

Implement shield platform encryption to encrypt files.

B.

Use classic encryption to encrypt files.

C.

Implement 3rd party App Exchange app to encrypt files.

D.

Store files outside of salesforce and access them to real time.

UC has a roll-up summary field on Account to calculate the count of contacts associated with an account. During the account load, SF is throwing an “Unable to lock a row” error.

Which solution should a data architect recommend, to resolve the error?

A.

Leverage data loader platform API to load data.

B.

Perform Batch job in parallel mode and reduce Batch size

C.

Perform Batch job in serial mode and reduce batch size

D.

Defer roll-up summary fields calculation during data migration.

UC has millions of Cases and are running out of storage. Some user groups need to have access to historical cases for up to 7 years.

Which 2 solutions should a data architect recommend in order to minimize performance and storage issues?

Choose 2 answers:

A.

Export data out of salesforce and store in Flat files on external system.

B.

Create a custom object to store case history and run reports on it.

C.

Leverage on premise data archival and build integration to view archived data.

D.

Leverage big object to archive case data and lightning components to show archived data.

Cloud Kicks is launching a Partner Community, which will allow users to register shipment requests that are then processed by Cloud Kicks employees. Shipment requests contain header information, and then a list of no more than 5 items being shipped.

First, Cloud Kicks will introduce its community to 6,000 customers in North America, and then to 24,000 customers worldwide within the next two years. Cloud Kicks expects 12 shipment requests per week per customer, on average, and wants customers to be able to view up to three years of shipment requests and use Salesforce reports.

What is the recommended solution for the Cloud Kicks Data Architect to address the requirements?

A.

Create an external custom object to track shipment requests and a child external object to track shipment items. External objects are stored off-platform in Heroku’s Postgres database.

B.

Create an external custom object to track shipment requests with five lookup custom fields for each item being shipped. External objects are stored off-platform in Heroku’s Postgres database.

C.

Create a custom object to track shipment requests and a child custom object to track shipment items. Implement an archiving process that moves data off-platform after three years.

D.

Create a custom object to track shipment requests with five lookup custom fields for each item being shipped Implement an archiving process that moves data off-platform after three years.