Page 1 of 2

One DataStage Server is going to die

Posted: Mon Jan 22, 2007 11:57 am
by fmartinsferreira
I have 2 DataStage Servers, Server1 and Server2.

Server1 needs to die :shock: and I need to put and make work all the Server1's project into Server2.

I'm comparing TNS, DB2 Catalog, tree of directories, ...

I know that to work they need to have their uvconfig file changed.

Today both work, in separate machines but I'm afraid that don't work together in the same machine. So I would like to know what is the best way to make this.

I know that there is a lot of parameters in uvconfig file but Let's supose that:

If Server1's uvconfig has "MFILES=100" and Server2's uvconfig has "MFILES=150" and like Server1 is goiing to die have I need to put "MFILES=250" in Server2's uvconfig file to work?

Regards,

Fernando Martins

Posted: Mon Jan 22, 2007 12:23 pm
by kcbland
No, you'll be fine. You're worried that having more jobs on one server is going to require higher settings. The MFILES is fine where you have it, just make sure your T30FILES is above 500 to support lots of simultaneous executing jobs.

Posted: Mon Jan 22, 2007 1:53 pm
by DSguru2B
You post heading is so tragic :shock:

Posted: Mon Jan 22, 2007 3:33 pm
by ray.wurlod
... unless you have an Australian accent, in which case "to die" is the day between yesterday and tomorrow. :lol:

Posted: Tue Jan 23, 2007 11:23 am
by fmartinsferreira
DSguru2B wrote:You post heading is so tragic :shock: ...
Sorry :)

What is the criteria to set DISKCACHE and DCBLOCKSIZE?

# DISKCACHE - Specifies the state of the DISKCACHE subsystem.
# Valid values are the following:
# 0 = REJECT, DISKCACHE is inactive
# and files opened in READONLY or WRITECACHE mode
# will give an error.
# -1 = ALLOW, the default value, DISKCACHE is inactive
# and files opened in READONLY or WRITECACHE mode
# are processed as if opened in READWRITE mode.
# n = DISKCACHE is active, where n is the size of the
# DISKCACHE shared memory in megabytes, and values
# 1-1000 are allowed.

Today I have DISKCACHE = 512! Why?

# DCBLOCKSIZE - Specifies the size of a DISKCACHE buffer
# in 1K units (1024 bytes).
# Valid values are 4, 8, 16, 32, and 64 with a default value of 16.

When DCBLOCKSIZE=16 work better than DCBLOCKSIZE=64?

It is the first time that I'm looking these parameters and I'm having a lot of doubts.

Fernando

Posted: Tue Jan 23, 2007 11:36 am
by ArndW
Fernando,

the settings are tunables for when you have the DataStage disk cacheing enabled. By setting the DISKCACHE value your system is now enabled for public hashed file cacheing. This does not necessarily mean that you are using the functionality, as a number of conditions and job settings need to be undertaken to start this.

The PDF describing this is on your client PC, in the DataStage directory under \Docs\dsdskche.pdf

Posted: Tue Jan 23, 2007 3:28 pm
by ray.wurlod
The disk cache tunables are only for hashed files, and therefore are irrelevant in the context of parallel jobs.

Posted: Tue Jan 23, 2007 5:34 pm
by vmcburney
Your main consideration is how the system will perform if you try and run the load of both servers on one server at the same time. For parallel jobs this means a lot of extra disk I/O, especially on the scratch space and temp space, do you have enough disk space allocated? Even if you run both sets of jobs at different times they may leave behind a lot of datasets that will take up disk space.

The other consideration is that if you overload a parallel system with too many jobs you can get unexpected aborts or slow performance due to lack of resources. A performance test will verify whether you are better off running server 1 and server 2 loads separately or whether they are efficient when running at the same time.

You can increase both the RAM and disk space on the machine if you have concerns.

Posted: Wed Jan 24, 2007 6:21 am
by fmartinsferreira
vmcburney wrote:Your main consideration is how the system will perform if you try and run the load of both servers on one server at the same time. For parallel jobs this means a lot of extra disk I/O, especially on the scratch space and temp space, do you have enough disk space allocated? Even if you run both sets of jobs at different times they may leave behind a lot of datasets that will take up disk space.

The other consideration is that if you overload a parallel system with too many jobs you can get unexpected aborts or slow performance due to lack of resources. A performance test will verify whether you are better off running server 1 and server 2 loads separately or whether they are efficient when running at the same time.

You can increase both the RAM and disk space on the machine if you have concerns.
The server that will continue to live is a very powerful machine.
RAM and disk space are not the problem. :D

When you said "scratch space and temp space" do you mean
SCRMAX and SCRSIZE parameters in uvconfig file?

Fernando

Posted: Wed Jan 24, 2007 1:32 pm
by ray.wurlod
No - these are for server jobs only. Scratch space is determined by the directories mentioned in the configuration file. Temporary space is determined by the directory mentioned in the TMPDIR environment variable.

Posted: Wed Jan 24, 2007 5:01 pm
by vmcburney
Some sorting and aggregation may go into your temp directory. Most datasets will be saved onto your node directories as defined in your configuration file. Your machine may have a lot of disk space but you need to make sure it is allocated correctly. Disk space monitoring during a test run will tell you what you need to know.

Posted: Mon Jan 29, 2007 1:19 pm
by fmartinsferreira
ray.wurlod wrote:No - these are for server jobs only. Scratch space is determined by the directories mentioned in the configuration file. Temporary space is determined by the directory mentioned in the TMPDIR enviro ...
Ray, :oops: sorry but:

Does the "configuration file" = uvconfig?

Where can I check how big is my Scratch space?

My TMPDIR enviroment variable is blank.

Fernando

Posted: Mon Jan 29, 2007 2:02 pm
by ArndW
PX sizing = look at your apt_config file for paths used
Server sizing = as mentioned earlier, look at the paths in uvconfig

Posted: Mon Jan 29, 2007 4:02 pm
by ray.wurlod
Configuration file is a file, probably in $DSHOME/../Configurations, whose name ends in ".apt" - for example default.apt

The current value of the APT_CONFIG_FILE environment variable determines which configuration file is in use.

Every parallel job that runs logs a message indicating which configuration file was used, and its contents. From this you can determine the paths of the directories used for disk and scratchdisk resource. From those you can determine the size of your scratch space.

If TMPDIR is empty, /tmp is used (or \tmp, if it exists, on the current drive on Windows).

Posted: Tue Jan 30, 2007 12:46 pm
by fmartinsferreira
Thanks! 8)

Fernando