CDC performance

Post questions here relative to DataStage Enterprise/PX Edition for such areas as Parallel job design, Parallel datasets, BuildOps, Wrappers, etc.

Moderators: chulett, rschirm, roy

Post Reply
adis
Participant
Posts: 13
Joined: Tue Feb 13, 2007 11:17 am

CDC performance

Post by adis »

Hi,
I have used CDC a couple of times on dimension tables with low volume of data(about 1 million records) and the performance was good.But now I have a requirement where I have to compare an entire record(20 columns) and the volume of data is high(21 million records).The job is running for the last two hours.Does the CDC stage support such a high volume of data? Does the "minus" operation in Oracle prove to be more efficient.Please suggest!Thanks so much!

Note:The volume of data keeps on growing for every load

Regards,
Adi
DSguru2B
Charter Member
Charter Member
Posts: 6854
Joined: Wed Feb 09, 2005 3:44 pm
Location: Houston, TX

Post by DSguru2B »

Minus is efficient but then again for a rather small dataset. How much is the byte size per row, that way you can calculate the amount of data you will be CDC'ing. Do test cases using both CDC stage and a simple minus function. Let us know your results.
Creativity is allowing yourself to make mistakes. Art is knowing which ones to keep.
Post Reply