I have two input files here,
Before_file1 & After_file2(sequential file stages).
Now I want to compare these two files and to get the MODIFIED, NEW-INSERTED & DELETED records.
I know this problem in Parallel Jobs, by using Change Caprute stage.
But I dont know in server addition.
In my current project it is necessary..
Can anyone help me?
Thanks in Advance
Comparing Two Datasets
Moderators: chulett, rschirm, roy
Load one of the files into a hashed file, the use that hashed file as a reference for the other. If the lookup fails it is a "N"ew one, if it succeeds then it is a "M"odified one.
Capturing deleted records this way is a bit more difficult, you could always load both files into hashed files and do lookups, or combine both files, sort and aggregate. A lot depends upon the sizes of your files.
Capturing deleted records this way is a bit more difficult, you could always load both files into hashed files and do lookups, or combine both files, sort and aggregate. A lot depends upon the sizes of your files.
<a href=http://www.worldcommunitygrid.org/team/ ... TZ9H4CGVP1 target="WCGWin">
</a>
</a>
Depending on how much data is involved with each row in the sequential files, follow Arnd's suggestion about using the hashed file, but only put the PK fields in there, this will make your hashed file smaller and possibly speed things up as well. Once again, this depends on the amount of data you are dealing with. Just a slight modification to his suggestion.
Hope that helps!
Hope that helps!
Bestest!
John Miceli
System Specialist, MCP, MCDBA
Berkley Technology Services
"Good Morning. This is God. I will be handling all your problems today. I will not need your help. So have a great day!"
John Miceli
System Specialist, MCP, MCDBA
Berkley Technology Services
"Good Morning. This is God. I will be handling all your problems today. I will not need your help. So have a great day!"
Depending on how much data is involved with each row in the sequential files, follow Arnd's suggestion about using the hashed file, but only put the PK fields in there, this will make your hashed file smaller and possibly speed things up as well. Once again, this depends on the amount of data you are dealing with. Just a slight modification to his suggestion.
Hope that helps!
Hope that helps!
Bestest!
John Miceli
System Specialist, MCP, MCDBA
Berkley Technology Services
"Good Morning. This is God. I will be handling all your problems today. I will not need your help. So have a great day!"
John Miceli
System Specialist, MCP, MCDBA
Berkley Technology Services
"Good Morning. This is God. I will be handling all your problems today. I will not need your help. So have a great day!"