Resource & Scratch Disk

Post questions here relative to DataStage Enterprise/PX Edition for such areas as Parallel job design, Parallel datasets, BuildOps, Wrappers, etc.

Moderators: chulett, rschirm, roy

Post Reply
nagarjuna
Premium Member
Premium Member
Posts: 533
Joined: Fri Jun 27, 2008 9:11 pm
Location: Chicago

Resource & Scratch Disk

Post by nagarjuna »

Hello All -- I need your suggestion regarding a setting up of configuration file

We have a cluster environment ( 4 servers ) & I am setting the configuration file .I am wondering what would be the best option of pointing scratch & resource disk . We have an NFS for which all those 4 servers has access .Also , Local space is present which is accessible be corresponding server.

a) As datastage creates lots of tem files while sorting in the scratch disk , I am planning to point scratch disk to be local . Please let me know if it is good option ?
b) Do I need to point the resource disk to the NFS or local disks . Please suggest .

Thanks
Nag
Nag
jwiles
Premium Member
Premium Member
Posts: 1274
Joined: Sun Nov 14, 2004 8:50 pm
Contact:

Post by jwiles »

1) Local storage for your scratch disks is recommended. SAN would be second depending upon it's performance. NFS is WAY down on the list for scratch storage

2) NFS would be better as far as easy access from all nodes of your cluster, with the connection to NFS on a private network separate (shared with nothing else) if possible and a minimum 1GB speed, 10GB would be better :)

Regards,
- james wiles


All generalizations are false, including this one - Mark Twain.
jassu
Premium Member
Premium Member
Posts: 7
Joined: Wed Dec 03, 2008 4:13 am
Location: USA

Post by jassu »

Thanks James for your response . For point 2 , Is there any drawback if we are mentioning local disk for resource disk ?
jwiles
Premium Member
Premium Member
Posts: 1274
Joined: Sun Nov 14, 2004 8:50 pm
Contact:

Post by jwiles »

Each server will have direct access to only the datasets (or portions thereof) stored on their local disks. This will make it difficult to run jobs on any available server (but not entirely impossible), hampering the ability to effectively balance processing loads across the cluster.

Regards,
- james wiles


All generalizations are false, including this one - Mark Twain.
ray.wurlod
Participant
Posts: 54607
Joined: Wed Oct 23, 2002 10:52 pm
Location: Sydney, Australia
Contact:

Post by ray.wurlod »

(What James said)... and, should you ever need to re-partition, you will need to move rows across the network.
IBM Software Services Group
Any contribution to this forum is my own opinion and does not necessarily reflect any position that IBM may hold.
nagarjuna
Premium Member
Premium Member
Posts: 533
Joined: Fri Jun 27, 2008 9:11 pm
Location: Chicago

Post by nagarjuna »

In case of repartition of the data , There is a need for each server to have an access for the all the resource disks specified . Let us suppose if I am specifying configuration file :

Node1 --- Server1 -- resource disk pointed local disk of server 1
Node2 --- Server2 -- resource disk pointed local disk of server 2
Node3 --- Server3 -- resource disk pointed local disk of server 3
Node4 --- Server4 -- resource disk pointed local disk of server 4

So , In this case will the job aborts in case of repartition ?
Nag
ray.wurlod
Participant
Posts: 54607
Joined: Wed Oct 23, 2002 10:52 pm
Location: Sydney, Australia
Contact:

Post by ray.wurlod »

No. Communications are established between the relevant player processes (the operators on each end of the link in which repartitioning is occurring) and they pass data to other nodes using APT_Communicator class (if I remember correctly) using TCP port numbers beginning at 11000 by default (again if I remember correctly).
IBM Software Services Group
Any contribution to this forum is my own opinion and does not necessarily reflect any position that IBM may hold.
nagarjuna
Premium Member
Premium Member
Posts: 533
Joined: Fri Jun 27, 2008 9:11 pm
Location: Chicago

Post by nagarjuna »

Thanks Ray & James for helping me out & clarifying this.

Resolved :D
Nag
Post Reply