Hashed File Creation Bogging Down Server
Moderators: chulett, rschirm, roy
-
- Premium Member
- Posts: 24
- Joined: Mon Aug 20, 2007 1:17 pm
Hashed File Creation Bogging Down Server
We have a UNIX server split into zones (all on Solaris 10). On one zone is DataStage, on the other is our reporting application. I've found that when creating large hashed files (>20 MB or so), the performance on the entire server suffers (the reporting app slows down, it takes significantly longer for Director to list the jobs in a category, and even doing a simple 'pwd' on the server takes 5-10 seconds instead of being instantaneous).
Judging by the server statistics, it appears to be a disk I/O limitation. We are willing to upgrade the hardware, but are wondering how much it needs to be upgraded.
Is there some kind of benchmark we can use to figure out how many bytes per second need to be written to the disk without crippling the server?
Also, is there some way to limit how much DataStage can write at one time so that performance on the rest of the server is not impacted?
Thanks for any insight that can be provided.
Judging by the server statistics, it appears to be a disk I/O limitation. We are willing to upgrade the hardware, but are wondering how much it needs to be upgraded.
Is there some kind of benchmark we can use to figure out how many bytes per second need to be written to the disk without crippling the server?
Also, is there some way to limit how much DataStage can write at one time so that performance on the rest of the server is not impacted?
Thanks for any insight that can be provided.
-
- Participant
- Posts: 54607
- Joined: Wed Oct 23, 2002 10:52 pm
- Location: Sydney, Australia
- Contact:
Much of this is automatic table space management within the dynamic hashed file structure. When you create the hashed file (as opposed to populating it) pre-size it according to the expected data volume - you can use Hashed File Calculator to achieve the calculation, and the MINIMUM.MODULUS keyword/option in the creation dialog.
A static hashed file of the correct size would be even better, but these are difficult to keep well tuned if the volume of data varies.
A static hashed file of the correct size would be even better, but these are difficult to keep well tuned if the volume of data varies.
IBM Software Services Group
Any contribution to this forum is my own opinion and does not necessarily reflect any position that IBM may hold.
Any contribution to this forum is my own opinion and does not necessarily reflect any position that IBM may hold.
-
- Premium Member
- Posts: 24
- Joined: Mon Aug 20, 2007 1:17 pm
I've seen something like this. At my last client's site, one particular job that populated a rather large hashed file would bring the server to its knees when it ran, right down to that same painful 15 second pwd. This wasn't a generic issue but was (thankfully) isolated to this one particular job. This was a Linux system with a fairly crappy disk subsystem but we never really did find out the root cause of this - DataStage, Linux, disk system, some combination... no clue.
Not really all that helpful for you but it turned out our hashed file was no longer needed by the process it was built to support so our ultimate solution was to... stop running the job.
Not really all that helpful for you but it turned out our hashed file was no longer needed by the process it was built to support so our ultimate solution was to... stop running the job.
-craig
"You can never have too many knives" -- Logan Nine Fingers
"You can never have too many knives" -- Logan Nine Fingers
-
- Premium Member
- Posts: 24
- Joined: Mon Aug 20, 2007 1:17 pm
Overrated? Are you kidding? They are the heart and soul of server jobs, something you will (and should be) leveraging to the hilt in whatever Server based solutions you build. Learn them. Love them.casedwgroup wrote:Hashed files are overrated anyway, right?
My issue was an aberration. Yours... disk, I assume, but not being a hardware guy can't really speak to the gory details. Have you had your SysAdmins monitor the system while the problematic hashed writes are happening? When I've had systems where 'tuning' of writes needed to be done, that could provide invaluable information they could use to improve performance. May be as simple as making config changes to how the 'journaling' functions (if such a thing applies) or it may be something it takes better hardware to improve.
-craig
"You can never have too many knives" -- Logan Nine Fingers
"You can never have too many knives" -- Logan Nine Fingers
-
- Premium Member
- Posts: 24
- Joined: Mon Aug 20, 2007 1:17 pm
-
- Participant
- Posts: 54607
- Joined: Wed Oct 23, 2002 10:52 pm
- Location: Sydney, Australia
- Contact:
-
- Premium Member
- Posts: 24
- Joined: Mon Aug 20, 2007 1:17 pm
-
- Premium Member
- Posts: 24
- Joined: Mon Aug 20, 2007 1:17 pm
-
- Participant
- Posts: 54607
- Joined: Wed Oct 23, 2002 10:52 pm
- Location: Sydney, Australia
- Contact: