Following are just guesses so don't read too much into them.
Check memory pool allocations, machine pool size, QAQQINI settings, system
value differences... Use WRKSYSSTS at the time it's happening and check
page faulting levels.
SMP shouldn't play a role since it's single processor box but perhaps having
it installed and turned-on on one box vs the other make query optimizer
behave differently.
Size of write cache on disks and otherwise differently performing drives
will make a difference (number of arms, data balanced etc.).
Amount of implicit journaling may play a role - SMAPP. System turns on
implicit journaling for large access paths so that may play a role from
resource conflict standpoint. Use EDTRCYAP to check that setting.
All that said, I'm guessing CPU is the bottleneck and just the fact that
there is other work going on concurrently and competing for CPU is causing
the bottleneck.
HTH, Elvis
Celebrating 10-Years of SQL Performance Excellence
http://centerfieldtechnology.com/training.asp
-----Original Message-----
Subject: SPAM-BIGDOG-- EDTRBDAP Question
I have a job that runs every night that effectively replicates a library
from our production box (810 - P10 processor with 8 gb ram) to our test
box (270 - P10 processor with 4 gb ram). Both machines are running V5R4
and are current with ptfs.
The job works like this.
10:30 pm on production: Six files are copied into a library named
BKPyymmdd. When this finishes the BKPyymmdd library is sent to the test
system using savrstlib. No problems here. When this finishes the identical
BKPyymmdd library exists on both systems.
2:00 am (after backups of both systems run at midnight), an identical job
starts on both systems that populates the 'audit' library on each system
using the six data files from the BKPyymmdd library. This process first
does a clrpfm on the six data files in the 'audit' library, and then does
a 'chglf xxfile MAINT(*REBLD)' for all of the logical files/sql indexes
that exist over these six physical files - (there are 23
logicals/indexes). Then cpyf is done (starting at RRN 1) to repopulate the
files, and when all of the cpyf's finish 'chglf xxfile MAINT(*IMMED)' is
done to rebuild the indexes. This method seems to be the fastest method
I've developled for completing this task overnight. The files are rather
large, three of them have over 25 million records (and there are no
deleted records).
The question is this: While this job runs on both sytems simultaneously,
when I monitor the access path rebuilds via EDTRBDAP, there is a column
that shows the estimated time to rebuild the access paths, but the funny
thing is that on our production system the estimated time is 30 to 50
percent greater than the estimated time for the same indexes on the test
system.
Our production system has double the ram and more cpw (810 vs 270). So why
would the rebuild of these access paths finish faster on a small box with
less horsepower? Granted there are a few other jobs on the production
system that do run overnight but typically these are queued and last night
(4:30 this morning when I checked it) there were only two other batch jobs
running. The test system is also our Domino Email Server so it's not
always idle overnight either. The production system also has more
available/free diskspace too.
I expect the usual answer - "It Depends"... but am looking for other clues
as to what could be causing this type of behavior, that results in the
longer access path rebuild time on production.
Anyone with any thoughts?
Regards, Jerry
As an Amazon Associate we earn from qualifying purchases.