Page 1 of 1

Converting Server Jobs to Parallel Jobs

Posted: Wed Jun 09, 2004 12:25 am
by Sreenivasulu
Hi All,

We are converting server jobs to parallel jobs. The trasformer
stage in the Canvas for Parallel Extender does not accept more than one input link (not even reference links). This is a big drawback since we cannot use the lookups(as reference links) from different hash files.

Regards

Posted: Wed Jun 09, 2004 1:49 am
by ray.wurlod
Parallel jobs require a completely different mind set from server jobs.
Read through the first section of the Parallel Job Developer's Guide to get a feel for what is needed.

Forget techniques you used in server jobs. Learn how to do it with parallel jobs. There is no hashed file available. There are three different stage types that perform "reference lookups". Everything works on in-memory datasets, so there's no need for hashed files. Datasets can be partitioned over the processing nodes described in the configuration file associated with your parallel job. You can specify different configuration files for different jobs, to organize different sets of resources for those jobs.

Posted: Wed Jun 09, 2004 5:21 am
by richdhan
Hi Sreenivas,

You can probably encapsulate your server job in a shared container and use it in your parallel job.

In one of the posts of Kenneth Bland I read that he was using parallel jobs for Extraction and Loading pocesses and server jobs for Transformation process. You can follow the same methodology.

Regards
Rich

A little bit of ink is powerful than the strongest memory
--Confucius

Posted: Wed Jun 09, 2004 4:52 pm
by ray.wurlod
There are several restrictions on when you can encapsulate server jobs in shared containers to use in parallel jobs - and let me note in passing that this is not the same as converting server jobs to parallel jobs.

The restrictions are documented on pages 2-19 and 2-20 of the version 7 DataStage Enterprise Edition Parallel Job Developer's Guide.