Page 1 of 1

Data set stage question

Posted: Fri Jul 06, 2007 12:49 pm
by jreddy
I have 3 independent jobs
1) Reads source data into dataset1 and existing target data into dataset2
2) The 2 above datasets are input to a change capture stage and the differences captured to a dataset3
3)Read dataset3 and apply changes via a transformer/filter

When i run these 3 jobs independently one after the other from designer, the datasets have the right information (datasets have OVERWRITE option)
But when i run them from sequencer, the datasets are not updated with correct/latest information, but seem to have data from prior runs and its producing incorrect results... What could be the possible problem?

i have made sure that job params are exactly the same and mapped correctly and also using 'DAtaset management' tool, i was able to see that the file modified time doesnt change when i run the jobs via sequencer..

Posted: Fri Jul 06, 2007 1:12 pm
by ray.wurlod
Are there any warnings in the log, particularly about changes to structure of Data Set, or "node not found", either if which might indicate use of an incompatible configuration file than that used when the Data Set was created.

Re: Data set stage question

Posted: Sat Jul 07, 2007 1:45 am
by ArndW
jreddy wrote:...But when i run them from sequencer, the datasets are not updated with correct/latest information...
What is your timing in the sequencer, you should have the completion of the first job trigger the next; it seem that you are not doing this and thus might be getting a copy of the old datasets before they are re-created.

Posted: Thu Jul 12, 2007 12:04 pm
by jreddy
There are no warnings in the log for the sequencer or the job.. it seems like they run fine, but just with the incorrect data in the data sets.. Also the timing is the same ballpark range when they would run individually..

I am wondering if there is some environment variable that needs to be set so we can share the data sets across jobs when run as part of sequencer..?

Posted: Thu Jul 12, 2007 7:24 pm
by ArndW
The reason I asked about timing was to make you check whether the dataset write job is complete before the dataset read job starts. You haven't confirmed or refuted that possible cause yet.

Posted: Fri Jul 13, 2007 9:17 am
by jreddy
In the sequencer, i have the triggers set to start the next job only when the first job is finished OK

the first job is to create the Datasets and second is to read those Datasets and do the CDC, so that makes me think the first job is done writing to the DS and then only the second job kicks in..

Posted: Fri Jul 13, 2007 3:28 pm
by ray.wurlod
To clarify, this only occurs when run from a job sequence? Try interposing an Execute Command or Routine activity between each pair of Job activities in which you will get the job sequence to sleep for, say, 30 seconds. This will give the file system a chance to flush the Data Sets to disk.