Jobs failing when running parallely

Post questions here relative to DataStage Enterprise/PX Edition for such areas as Parallel job design, Parallel datasets, BuildOps, Wrappers, etc.

Moderators: chulett, rschirm, roy

Post Reply
pradeep9081
Participant
Posts: 30
Joined: Tue May 11, 2010 2:05 pm

Jobs failing when running parallely

Post by pradeep9081 »

Hi,

We are running multiple jobs (5-6 jobs) at a time from the scheuduler.
Jobs are failing and getting the below error:

Unable to start ORCHESTRATE process on node node1 (nsyrp41b): APT_PMPlayer::APT_PMPlayer: fork() failed, Not enough space.

If i run the individual job then works fine.

If there is a look up then getting the below error:

"/eeadm2/IBM/InformationServer/Server/Datasets/lookuptable.20100803.jm02qdd": No space left on device
APT_BufferOperator: Add block to queue failed. This means that your buffer file systems all ran out of file space, or that some other system error occurred. Please ensure that you have sufficient scratchdisks in either the default or "buffer" pools on all nodes in your configuration file.

we have 2 node configuration file in dev. pointing both nodes to same scratch disk space.

is this dues to the buffer size? what is the best resolution for this ?
kris007
Charter Member
Charter Member
Posts: 1102
Joined: Tue Jan 24, 2006 5:38 pm
Location: Riverside, RI

Re: Jobs failing when running parallely

Post by kris007 »

pradeep9081 wrote: "/eeadm2/IBM/InformationServer/Server/Datasets/lookuptable.20100803.jm02qdd": No space left on device
APT_BufferOperator: Add block to queue failed. This means that your buffer file systems all ran out of file space, or that some other system error occurred. Please ensure that you have sufficient scratchdisks in either the default or "buffer" pools on all nodes in your configuration file.

we have 2 node configuration file in dev. pointing both nodes to same scratch disk space.

is this dues to the buffer size? what is the best resolution for this ?
Yes. The error message says it all. Your scratch disk space is full. You need to add extra space or reschedule your jobs so that they don't run at the same time.
Kris

Where's the "Any" key?-Homer Simpson
mouthou
Participant
Posts: 208
Joined: Sun Jul 04, 2004 11:57 pm

Re: Jobs failing when running parallely

Post by mouthou »

Or if the diskspace management is out of your control, the job(s) can be modified to use Join stage instead of Lookup or use the lookup with the filter conditions :idea:
Barath
Participant
Posts: 17
Joined: Mon Sep 29, 2008 4:00 am
Location: Mumbai

Re: Jobs failing when running parallely

Post by Barath »

pradeep9081 wrote:Hi,

We are running multiple jobs (5-6 jobs) at a time from the scheuduler.
Jobs are failing and getting the below error:

Unable to start ORCHESTRATE process on node node1 (nsyrp41b): APT_PMPlayer::APT_PMPlayer: fork() failed, Not enough space.

If i run the individual job then works fine.

If there is a look up then getting the below error:

"/eeadm2/IBM/InformationServer/Server/Datasets/lookuptable.20100803.jm02qdd": No space left on device
APT_BufferOperator: Add block to queue failed. This means that your buffer file systems all ran out of file space, or that some other system error occurred. Please ensure that you have sufficient scratchdisks in either the default or "buffer" pools on all nodes in your configuration file.

we have 2 node configuration file in dev. pointing both nodes to same scratch disk space.

is this dues to the buffer size? what is the best resolution for this ?
Which Partitioning you are using . If it is Entire then change it to Hash.
If you are still getting issue with space then you need to add space what kris007 is saying correct.
Post Reply