Job fails

Post questions here relative to DataStage Enterprise/PX Edition for such areas as Parallel job design, Parallel datasets, BuildOps, Wrappers, etc.

Moderators: chulett, rschirm, roy

Post Reply
Nagaraj
Premium Member
Premium Member
Posts: 383
Joined: Thu Nov 08, 2007 12:32 am
Location: Bangalore

Job fails

Post by Nagaraj »

i have a pretty simple design

oracle connector----> TFM--->LKP--->dataset

the source is based on 3 views joined together, rows are about 5 million.
trying to run the job on default config first i.e 1 node.

The job fails with the following message.

APT_CombinedOperatorController,0: Write to dataset on [fd 8] failed (Success) on node node1,

APT_CombinedOperatorController,0: Orchestrate was unable to write to any of the following files:

APT_CombinedOperatorController,0: /IBM/InformationServer/Server/Datasets/ctproceduredelted.ds.xxx.hostname.0000.0000.0000.2d9b.cf8abf34.0001.368bc84f

APT_CombinedOperatorController,0: Block write failure. Partition: 0

Any idea where do i start first to look into ???
Nagaraj
Premium Member
Premium Member
Posts: 383
Joined: Thu Nov 08, 2007 12:32 am
Location: Bangalore

Post by Nagaraj »

I have checked the disck space everything looks fine to me,
the disk on which DS in installed is only 30% filled
and the data files directory is only 35% filled.

so its clearly not an issue with the disk space.

and also the ulimit -a is set to unlimited and i am running on 32 bit linux OS.
Nagaraj
Premium Member
Premium Member
Posts: 383
Joined: Thu Nov 08, 2007 12:32 am
Location: Bangalore

Post by Nagaraj »

Also changed the LKp to reference Datasets instead of Database, still the issue exists.
chulett
Charter Member
Charter Member
Posts: 43085
Joined: Tue Nov 12, 2002 4:34 pm
Location: Denver, CO

Post by chulett »

You checked the disk space available while the job was running?
-craig

"You can never have too many knives" -- Logan Nine Fingers
Nagaraj
Premium Member
Premium Member
Posts: 383
Joined: Thu Nov 08, 2007 12:32 am
Location: Bangalore

Post by Nagaraj »

1.Tried with the Oracle EE stage, same error message appears,
also tried changing the paremeters, auto buffering mode etc.
2. Tried simplifying the Job to read from db and copy stage followed by Dataset.


all of them are sending out the same error message at the source stage itself.
Nagaraj
Premium Member
Premium Member
Posts: 383
Joined: Thu Nov 08, 2007 12:32 am
Location: Bangalore

Post by Nagaraj »

chulett wrote:You checked the disk space available while the job was running?
yes no change,
Nagaraj
Premium Member
Premium Member
Posts: 383
Joined: Thu Nov 08, 2007 12:32 am
Location: Bangalore

Post by Nagaraj »

chulett wrote:You checked the disk space available while the job was running?
checked again seems like the disk space growing close to 74% and still increasing, i killed the job and its aborted now, still it is increasing, any way to stop this before the server is shut down.
Nagaraj
Premium Member
Premium Member
Posts: 383
Joined: Thu Nov 08, 2007 12:32 am
Location: Bangalore

Post by Nagaraj »

what do we have here

/IBM/InformationServer/Server/Datasets

believe this is the tmp directory used for paging?
Nagaraj
Premium Member
Premium Member
Posts: 383
Joined: Thu Nov 08, 2007 12:32 am
Location: Bangalore

Post by Nagaraj »

/IBM/InformationServer/Server/Datasets
this is eating up all the space.

is there any way to give a different path instead of the above, if yes? where do i give it?
chulett
Charter Member
Charter Member
Posts: 43085
Joined: Tue Nov 12, 2002 4:34 pm
Location: Denver, CO

Post by chulett »

That would be controlled by your config file.
-craig

"You can never have too many knives" -- Logan Nine Fingers
Nagaraj
Premium Member
Premium Member
Posts: 383
Joined: Thu Nov 08, 2007 12:32 am
Location: Bangalore

Post by Nagaraj »

gr8 thanks, made the changes.
now i will focus on the issue. see if this problem araises again.
Post Reply