Whole Machine Grinds to a Halt - On Disk Contention.

Post questions here relative to DataStage Server Edition for such areas as Server job design, DS Basic, Routines, Job Sequences, etc.

Moderators: chulett, rschirm, roy

Post Reply
Jluchjen
Participant
Posts: 5
Joined: Wed Jul 28, 2004 2:29 am
Location: Darwin Australia
Contact:

Whole Machine Grinds to a Halt - On Disk Contention.

Post by Jluchjen »

We are Running:
Datastage Server Version 7.1r1
AIX version 5.2.
Disk is an Hitachi 9500 disk array
connected over Dual Channel a fibre links.

Hello All,

When I run a couple of my Jobs which generate a large amount of Disk IO. This AIX box performs like a dog, ie telnet and issue an ls you may wait up to 2 minutes for a response or try to exit a shell the same sort of wait.

My system administrator and I feel we are getting nowhere and recovering the same ground.

When we use nmon to watch what is happening, the following things are observed:

Once the hard disk reaches 100% (Stating anything from 3000kb to 4000kb per Second Through put) One by one the CPUs show 100% W (Wait). The IO appears to continue to be written, and when it clears the CPUs seem to free up again for a while. Only to repeat the Cycle.

Has anybody seen this before (and hopefully fixed it)????

Thanks Jack
kcbland
Participant
Posts: 5208
Joined: Wed Jan 15, 2003 8:56 am
Location: Lutz, FL
Contact:

Post by kcbland »

Yep, happens all the time. One simple job SEQ --> XFM --> HASH can hog a poor disk/controller design. Try running two jobs like that at the same time, forget about it. Put a database on top of that layout, go brew a pot of coffee and send out for donuts.

We see this on linux and Windows boxes a lot, the disk framework is not organized to handle the high disk i/o requirements. For example, one of my customers development server uses Clarions but the production server uses Symmetrix. The jobs on the development server can't execute a full volume data run because of disk contention, but the production server handles it without breaking a sweat.

Your disk array, can you describe it's nature? Is it a heck of a lot of disks and controllers like the high-end 9500, or are we talking 5 disks striped on a single controller with 1 GB cache? What kind of volume are you throwing at it?
Kenneth Bland

Rank: Sempai
Belt: First degree black
Fight name: Captain Hook
Signature knockout: right upper cut followed by left hook
Signature submission: Crucifix combined with leg triangle
Jluchjen
Participant
Posts: 5
Joined: Wed Jul 28, 2004 2:29 am
Location: Darwin Australia
Contact:

Post by Jluchjen »

Thanks Ken,

That was quick we are truly gratefull.

My Sysetm admin says:
Our 9500 has 4GB of cache.
4 raid controllers.
There is 13 raid5 groups of 4data + 1 parity.

We are defining 150GB disks within these raid groups.
AIX is connected over 2Gb fibre link.

Ta Jack
gpatton
Premium Member
Premium Member
Posts: 47
Joined: Mon Jan 05, 2004 8:21 am

Post by gpatton »

Had a very similar problem before.

What you need to do is see how the disk drives / LUN's are actually mapped to the RAID groups.

You may be trying to separate IO and in reality be exasperating it.

If Disk 1, Disk 2, Disk 3 and Disk4 are used in LUN1 and LUN2 and LUN3, then you should make all of these LUN's part of one file system.

Make sure that the database log files are on separate disk from the database files.

Also check out the RAID stripping on the SAN and the database to make sure that you are not cross stripping the IO.
Jluchjen
Participant
Posts: 5
Joined: Wed Jul 28, 2004 2:29 am
Location: Darwin Australia
Contact:

Post by Jluchjen »

Hi :D ,

Thank you very much, I have passed on your comments to our System Administrator and look forward to improved performance.

I intend to post again once we get the changes Implemented and measure the performance gain.

Ta Jack
Post Reply