Hi there
I want to match data between two data sources based on a common field. My issue is that I have a large data set of 200k-500k and there are duplicate records that I would like to keep in the dataset. Because of the volume the find and replace tool (I think) has issues with performance therefore always producing fluctuating numbers, whilst the join tool creates massive Cartesian Join creating millions of rows that I can use a unique tool to reduce back to a normal size but significantly increase the runtime.
I really want to use the find and replace tool but because of inconsistent results I can only rely on the join tool which is a pain in the * to use cause of the Cartesian Join. Can anyone suggest what I could do in this situation?