Hi Henrik,

unfortunately I don't have any specific recommendations for you situation, but 
consider looking at some of these:

(our ADSM server is a silver node, 4-way, 2 GB memory)

The following are segments from an article that I'm writing, and hoping to get 
published, if the information is bad or good please let me know!

1. example from my TSM (ADSM) server:
        The biggest benefit I found is tuning vmtune. It is found in 
/usr/samples/kernel/vmtune (if you have installed bos.adt.samples). The parameters 
that I have changed are:

        -r      'r' is the minimum number of pages that need to be read sequentially 
before the VMM detects a sequential read. I have this set to 2.

        -R      to find to right setting for R, I used /usr/bin/filemon. First I set R 
to an arbitrary large value, such as 128. 
                                Then I started a filemon session and copied a large 
sequential file to /dev/null. Next look at the filemon report for that LV contains the 
filesystem:
                example: from ADSM server node
                VOLUME: /dev/dbdumps_lv  description: /home/dbdumps
                reads:                  3356    (0 errs)
                  read sizes (blks):    avg   194.3 min       8 max     256 sdev    
90.6
                  read times (msec):    avg  35.693 min   1.096 max 488.857 sdev  
39.371
                  read sequences:       2835
                  read seq. lengths:    avg   230.0 min       8 max     512 sdev    
68.5

                Notice that the min read blocks is 8 and the max is 256. Remember a 
block is 512 bytes. So the max is actually 128k or 32 x 4k blocks. 
                               Which means the maximum number of read ahead blocks 
that will happen is 32. This number will be dependent on your adapter and your disks.

        -f      'f' is the lower limit of free pages. When the free list hits this 
number the VMM starts to free memory pages. IBM recommends that this number be: 
                                120 x the number of CPU's you have. I would try 
different combinations to see what gives you the best performance. Most of my boxes 
have this set at 256.

        -F      'F' is the upper limit of free pages. When the VMM is freeing pages 
and hits this many free it stops. 'F' should be at least: f + R x 4096). 
                                This guarantees that there will be enough free memory 
for a large disk read. I follow this for all of my boxes.

        -h      Setting 'h' to '1' (one) will enforce maxperm. I had to set this on my 
production nodes because they where using too much memory (8-10 GB) 
                                for file caching, managing this amount of file cache 
was causing the HACMP daemons to time out, causing HACMP to think the node had failed. 

                WARNING: DO NOT CHANGE THESE PARAMETERS ON THE FLY - especially on a 
production machine, it may have disastrous results. 
                                I recommend changing them in /etc/rc.local.

2. Queue depth
If you are using SSA disks and especially if you are using a RAID array check the 
queue depth.
EX:
root@unxd:/>lsattr -El hdisk5 
        pvid            00011396f62aa5420000000000000000 Physical volume identifier  
False
        queue_depth     112                              Queue depth                 
True
        write_queue_mod 1                                Write queue depth modifier  
True
        adapter_a       ssa0                             Adapter connection          
False
        adapter_b       none                             Adapter connection          
False
        primary_adapter adapter_a                        Primary adapter             
True
        reserve_lock    yes                              RESERVE device on open      
True
        connwhere_shad  139639D79D1F4CE                  SSA Connection Location     
False
        max_coalesce    0x20000                          Maximum coalesced operation 
True
        size_in_mb      255117                           Size in Megabytes           
False
        location                                         Location Label              
True

        If the last column is 'True' then the value can be changed, if it is 'False', 
then the information is for display only.

        One parameter that I always check, especially with RAID arrays, is the 
queue_depth. 
                This specifies the number of commands that can be outstanding against 
the logical disk. It is very important with RAID arrays to make sure that this number 
is: 
        the queue depth of an individual disk x the number of disks in the array. 
                In this example it is 8 x 14 or 112. It is not 16 disks because the 
equivalent of one disk is used for parity and one hot spare.

3. Are you using LVM striping?

4. Are you using the fast write cache (FWC)?

5. Is your TSM database on the same disks as your disk storage pool? You may want to 
separate them if they are.

6. other recommendations:
-put the busiest disks in the centre of the loop, to give the less active disks a 
chance to transfer information
-when using RAID 1 or RAID 10, make sure that half the disks are closest to one port 
and the other half of the disks are closest to the other port
-try not to put more than 32 disks per loop
-use the fast write cache, especially for RAID 5 array's

Our setup:
I have 16 SSA disks.

  - 2 disks are configured as one RAID 1 array, in positions 1 and 16 in the I/O 
drawer - equal distance from the adapter. This holds the ADSM database and log.
  - 6 disks hold my disk pool (RAID 0)
  - 8 disks hold a filesystem that is used to back Sybase databases. It is NFS mounted 
to all the other nodes.

Let me know if I can help.

Miles

-------------------------------------------------------------------------------------------
Miles Purdy 
System Manager
Farm Income Programs Directorate
Winnipeg, MB, CA
[EMAIL PROTECTED]
ph: (204) 984-1602 fax: (204) 983-7557
-------------------------------------------------------------------------------------------

>>> [EMAIL PROTECTED] 08-May-01 3:51:19 AM >>>
I need some advice on performance issues regarding tsm 4.1.2.0 running on
a aix 4.3.3 box.

I've been trying to optimize the network throughput, but without real
luck. Then I tried backing up the tsm server through lo0 and I ended up
with the following result

Total number of objects inspected:   49,295
Total number of objects backed up:   49,295
Total number of objects updated:          0
Total number of objects rebound:          0
Total number of objects deleted:          0
Total number of objects expired:          0
Total number of objects failed:           0
Total number of bytes transferred:     2.45 GB
Data transfer time:                  546.68 sec
Network data transfer rate:        4,706.32 KB/sec
Aggregate data transfer rate:      2,041.05 KB/sec
Objects compressed by:                    0%
Elapsed processing time:           00:03:00

which is not a fantastic throughput.

I'm backing up to a diskpool of ssa-disks (no raid) and i have a 30GB
database which has never been defragmented.

Is it a memory tuning issue?

Any suggestions are welcome.

Does anyone know of a new tivoli performance tuning guide? I have a (not
up to date) SP performance tuning guide.

Med venlig hilsen / Regards

Henrik Ursin            Tlf./Phone +45 35878934
                        Fax        +45 35878990
                        Email [EMAIL PROTECTED] 
                        Mail: UNI-C
                              DTU, bygning 304
                              DK-2800 Lyngby

Reply via email to