Skip to main content

Is there DSP documentation on bulk duplicate detection?

Answered

Comments

6 comments

  • Best answer
    DSP Expert

    CBermingham: Hi Paul, The documentation I have is the help. Go to Platform > Common > Find Duplicates Overview. There should be enough there to get you going. Regards, Chris

    0
  • DSP User

    Thank you Chris. I have used that documentation to initially set up the Dup checking process. However, I have not found any options for reporting on the results other than granting users security access to the Common/Analyze WebApp to view results. The client would like a report option. I could create these reports, but I wondered if someone has done this already. Also, the record sets at the client are large for customers and vendors. Support is suggesting a max of 40k. Have you found a way around this?

    0
  • DSP Expert

    CBermingham: I have been able to run with datasets as large as 50k records, but it depends on how many fields you're evaluating. To process a larger population, I had to split the data into sets that would be small enough, but would not separate likely duplicates. So with Vendors, I'd start with splitting by country on the belief that Acme in Ohio, USA is not going to merge with an Acme in Japan. Each slice gets processed as a separate duplicate object.Then you'd need to incorporate the results for each duplicate object into your master. Adjusting the thresholds is a matter of trial and error.

    0
  • DSP User

    Thank you, Chris.

    0
  • DSP Expert

    JGrippa: There is an update to DSP Help regarding dDup as of version 6.6.1. Hopefully, this will help a bit.

    0
  • DSP User

    Thank you, John Paul. This will be very helpful.

    0

Please sign in to leave a comment.