Here is what I was thinking:kris007 wrote:Hey Craig, I am doing exactly what Ken has suggested. But Raghu seems to come up with a different idea. Was just trying to findout what he is suggesting, clearly so that it would help. It wouldn't hurt to know if there is another way..rightKris, that's basically what Ken explained in his post.![]()
Kris
1) You have 80 million records in Target with a timestamp.
2) You load these 80 million keys into a hash file (this is a problem process and takes lot of time according to you).
3)You get 1 million incremental records...you look up the hash file and you are either updating or inserting based on the availability of the keys. This is not a problem process according to you. So now you have (say) 80.6 million records in target , assuming .4 million records were updates.
4) Now try to insert the 1 million records into your hash file ie. do not clear the hash file before you insert, based on your timestamp. Your hash file will now have 80.6 million records too. This time this job will be quick as the no. of records is less.
Repeat step 3) and step 4) for your incremental loads..
Wouldn't this work ? Have I missed out something ?
However, this design is not complete and still has other problems like what happens when you want to do a full load ? You might have to handle it through other jobs.
Thanks,
Raghu