New Year Sale 2026! Hurry Up, Grab the Special Discount - Save 25% - Ends In 00:00:00 Coupon code: SAVE25
Welcome to Pass4Success

- Free Preparation Discussions

Salesforce Certified Platform Data Architect (Plat-Arch-201) Exam - Topic 1 Question 52 Discussion

Actual exam question for Salesforce's Salesforce Certified Platform Data Architect (Plat-Arch-201) exam
Question #: 52
Topic #: 1
[All Salesforce Certified Platform Data Architect (Plat-Arch-201) Questions]

Universal Containers has been a customer of Salesforce for 10 years. Currently they have 2 million accounts in the system. Due to an erroneous integration built 3 years ago, it is estimated there are 500,000 duplicates in the system.

Which solution should a data architect recommend to remediate the duplication issue?

Show Suggested Answer Hide Answer
Suggested Answer: D

Implementing duplicate rules (option D) is the best solution to remediate the duplication issue, as it allows the data architect to identify and merge duplicate accounts in Salesforce using native features and tools. Developing an ETL process that utilizes the merge API to merge the duplicate records (option A) is not a good solution, as it may require more coding and testing effort, and it does not prevent duplicates from being created in Salesforce. Utilizing a data warehouse as the system of truth (option B) is also not a good solution, as it may introduce additional complexity and cost, and it does not address the duplication issue in Salesforce. Extracting the data using data loader and using excel to merge the duplicate records (option C) is also not a good solution, as it may be time-consuming and error-prone, and it does not prevent duplicates from being created in Salesforce.


Contribute your Thoughts:

0/2000 characters
Emiko
3 months ago
Using a data warehouse sounds like overkill for this issue.
upvoted 0 times
...
Lili
3 months ago
Totally agree with you, Tom! ETL process is the way to go.
upvoted 0 times
...
Serina
3 months ago
Wait, can Excel really handle that many records?
upvoted 0 times
...
Shaquana
4 months ago
I think option A is the best way to handle this.
upvoted 0 times
...
Kimi
4 months ago
500,000 duplicates? That's a huge mess!
upvoted 0 times
...
Micah
4 months ago
I practiced a similar question where we used data loader and Excel, but that seems too manual for 500,000 duplicates.
upvoted 0 times
...
Kati
4 months ago
The ETL process sounds familiar, but I wonder if it would be too complex for this situation.
upvoted 0 times
...
Rhea
4 months ago
I think using duplicate rules could be a quick fix, but it might not address the existing duplicates effectively.
upvoted 0 times
...
Cristina
5 months ago
I remember we discussed the merge API in class, but I'm not entirely sure if it's the best option for such a large number of duplicates.
upvoted 0 times
...
Vallie
5 months ago
Implementing duplicate rules could be a good way to prevent future duplication, but it doesn't address the existing issue. I'd want to combine that with another solution.
upvoted 0 times
...
Charlena
5 months ago
Extracting the data and using Excel to merge the duplicates seems like a manual and error-prone approach. I'd want to avoid that if possible.
upvoted 0 times
...
Alberto
5 months ago
Utilizing a data warehouse as the system of truth could be a solid solution, but I'd need to understand how that would integrate with their existing Salesforce setup.
upvoted 0 times
...
Fabiola
5 months ago
Developing an ETL process to merge the duplicate records using the Merge API sounds like a good approach, but I'm not sure how efficient that would be for 500,000 duplicates.
upvoted 0 times
...
Alica
5 months ago
Hmm, this seems like a tricky one. I'll need to carefully consider the pros and cons of each option.
upvoted 0 times
...
Gladys
1 year ago
I'd go with option D. Implement duplicate rules and let Salesforce handle the deduplication automatically. Easy peasy!
upvoted 0 times
...
Alona
1 year ago
Haha, using a data warehouse as the 'system of truth' for Salesforce data? That's like using a sledgehammer to crack a nut. Option B is way overkill.
upvoted 0 times
Sophia
1 year ago
D) Implement duplicate rules
upvoted 0 times
...
Lorrine
1 year ago
C) Extract the data using data loader and use excel to merge the duplicate records
upvoted 0 times
...
Evangelina
1 year ago
B) Utilize a data warehouse as the system of truth
upvoted 0 times
...
Catarina
1 year ago
A) Develop an ETL process that utilizers the merge API to merge the duplicate records
upvoted 0 times
...
...
Cathern
1 year ago
Option C sounds too manual and time-consuming for a 2 million record dataset. Excel is not the right tool for this kind of large-scale data cleanup.
upvoted 0 times
Caprice
1 year ago
D) Implement duplicate rules
upvoted 0 times
...
Belen
1 year ago
B) Utilize a data warehouse as the system of truth
upvoted 0 times
...
Alline
1 year ago
A) Develop an ETL process that utilizers the merge API to merge the duplicate records
upvoted 0 times
...
...
Ty
1 year ago
But wouldn't using data loader and excel be a quicker fix?
upvoted 0 times
...
Kerrie
1 year ago
I think option A is the best choice. Using the Salesforce Merge API to automate the deduplication process is the most comprehensive solution here.
upvoted 0 times
...
Bambi
1 year ago
Definitely option D. Implementing duplicate rules is the way to go for a large dataset like this. It's a scalable and efficient solution.
upvoted 0 times
Lavelle
1 year ago
I'm leaning towards option D as well, it seems like the most practical choice.
upvoted 0 times
...
Afton
1 year ago
I would go with option B, using a data warehouse seems like a solid solution.
upvoted 0 times
...
Christoper
1 year ago
I think option A could also work well if implemented correctly.
upvoted 0 times
...
Hoa
1 year ago
I agree, option D is definitely the best choice for this situation.
upvoted 0 times
...
...
Leila
1 year ago
I disagree, I believe implementing duplicate rules would be a more efficient solution.
upvoted 0 times
...
Ty
1 year ago
I think we should develop an ETL process to merge the duplicates.
upvoted 0 times
...

Save Cancel