Re: [Veritas-vx] Solaris-SFS / MPxIO / VxVM failover issue

2010-09-16 Thread Joshua Fielden
dmp_fast_recovery is a mechanism by which we bypass the sd/scsi stack and send 
path inquiry/status CDBs directly from the HBA in order to bypass long SCSI 
queues and recover paths faster. With a TPD (third-party driver) such as MPxIO, 
bypassing the stack means we bypass the TPD completely, and interactions such 
as this can happen. The vxesd (event-source daemon) is another 5.0/MP2 backport 
addition that's moot in the presence of a TPD.

From your modinfo, you're not actually running MP3. This technote 
(http://seer.entsupport.symantec.com/docs/327057.htm) isn't exactly your 
scenario, but looking for partially-installed pkgs is a good start to getting 
your server correctly installed, then the tuneable should work -- very early 
5.0 versions had a differently-named tuneable I can't find in my mail archive 
ATM.

Cheers,

Jf

-Original Message-
From: veritas-vx-boun...@mailman.eng.auburn.edu 
[mailto:veritas-vx-boun...@mailman.eng.auburn.edu] On Behalf Of Sebastien 
DAUBIGNE
Sent: Thursday, September 16, 2010 7:41 AM
To: Veritas-vx@mailman.eng.auburn.edu
Subject: Re: [Veritas-vx] Solaris-SFS / MPxIO / VxVM failover issue

  Thank you Victor and William, it seems to be a very good lead.

Unfortunately, this tunable seems not to be supported in the VxVM 
version installed on my system :

  vxdmpadm gettune dmp_fast_recovery
VxVM vxdmpadm ERROR V-5-1-12015  Incorrect tunable
vxdmpadm gettune [tunable name]
Note - Tunable name can be dmp_failed_io_threshold, dmp_retry_count, 
dmp_pathswitch_blks_shift, dmp_queue_depth, dmp_cache_open, 
dmp_daemon_count, dmp_scsi_timeout, dmp_delayq_interval, dmp_path_age, 
or dmp_stat_interval

Something odd because my version is 5.0 MP3 Solaris SPARC, and according 
to http://seer.entsupport.symantec.com/docs/316981.htm this tunable 
should be available.

  modinfo | grep -i vx
  38 7846a000  3800e 288   1  vxdmp (VxVM 5.0-2006-05-11a: DMP Drive)
  40 784a4000 334c40 289   1  vxio (VxVM 5.0-2006-05-11a I/O driver)
  42 783ec71ddf8 290   1  vxspec (VxVM 5.0-2006-05-11a control/st)
296 78cfb0a2c6b 291   1  vxportal (VxFS 5.0_REV-5.0A55_sol portal )
297 78d6c000 1b9d4f   8   1  vxfs (VxFS 5.0_REV-5.0A55_sol SunOS 5)
298 78f18000   a270 292   1  fdd (VxQIO 5.0_REV-5.0A55_sol Quick )





Le 16/09/2010 12:15, Victor Engle a écrit :
 Which version of veritas? Version 4/2MP2 and version 5.x introduced a
 feature called DMP fast recovery. It was probably supposed to be
 called DMP fast fail but recovery sounds better. It is supposed to
 fail suspect paths more aggressively to speed up failover. But when
 you only have one vxvm DMP path, as is the case with MPxIO, and
 fast-recovery fails that path, then you're in trouble. In version 5.x,
 it is possible to disable this feature.

 Google DMP fast recovery.

 http://seer.entsupport.symantec.com/docs/307959.htm

 I can imagine there must have been some internal fights at symantec
 between product management and QA to get that feature released.

 Vic





 On Thu, Sep 16, 2010 at 6:03 AM, Sebastien DAUBIGNE
 sebastien.daubi...@atosorigin.com  wrote:
   Dear Vx-addicts,

 We encountered a failover issue on this configuration :

 - Solaris 9 HW 9/05
 - SUN SAN (SFS) 4.4.15
 - Emulex with SUN generic driver (emlx)
 - VxVM 5.0-2006-05-11a

 - storage on HP SAN (XP 24K).


 Multipathing is managed by MPxIO (not VxDMP) because the SAN team and HP
 support imposed the Solaris native solution for multipathing :

 VxVM ==  VxDMP ==  MPxIO ==  FCP ...

 We have 2 paths to the switch, linked to 2 paths to the storage, so the
 LUNs have 4 paths, with active/active support.
 Failover operation has been tested successfully by offlining each port
 successively on the SAN.

 We regulary have transient I/O errors (scsi timeout, I/O error retries
 with Unit attention), due to SAN-side issues. Usually these errors are
 transparently managed by MPxIO/VxVM without impact on the applications.

 Now for the incident we encountered :

 One of the SAN port was reset , consequently there were some transient
 I/O error.
 The other SAN port was OK, so the MPxIO multipathing layer should have
 failover the I/O on the other path, without transmiting the error to the
 VxDMP layer.
 For some reason, it did not failover the I/O before VxVM caught it as
 unrecoverable I/O error, disabling the subdisk and consequently the
 filesystem.

 Note the giving up message from scsi layer at 06:23:03 :

 Sep  1 06:18:54 myserver vxdmp: [ID 917986 kern.notice] NOTICE: VxVM
 vxdmp V-5-0-112 disabled path 118/0x558 belonging to the dmpnode 288/0x60
 Sep  1 06:18:54 myserver vxdmp: [ID 824220 kern.notice] NOTICE: VxVM
 vxdmp V-5-0-111 disabled dmpnode 288/0x60
 Sep  1 06:18:54 myserver vxdmp: [ID 917986 kern.notice] NOTICE: VxVM
 vxdmp V-5-0-112 disabled path 118/0x538 belonging to the dmpnode 288/0x20
 Sep  1 06:18:54 myserver vxdmp: [ID 917986 kern.notice] NOTICE: VxVM
 vxdmp V-5-0-112 disabled path 118/0x550 belonging to the dmpnode 288/0x18
 Sep  1 06:18:54 

Re: [Veritas-vx] Relayout volume from 2 to 4 columns

2010-01-06 Thread Joshua Fielden
Striping across Enterprise LUNs allows you to use more SCSI queues/transaction, 
and follow the template of setting the number of read/write threads to the 
number of stripe cols, and have it work efficiently, amongst other arguments 
for the config.

Not everyone needs this configuration, but it's the difference between made and 
missed SLAs for a lot of configurations.

Cheers,

Jf

-- 
Sent using BlackBerry 




From: veritas-vx-boun...@mailman.eng.auburn.edu 
veritas-vx-boun...@mailman.eng.auburn.edu 
To: veritas-vx@mailman.eng.auburn.edu veritas-vx@mailman.eng.auburn.edu 
Sent: Wed Jan 06 10:19:56 2010
Subject: Re: [Veritas-vx] Relayout volume from 2 to 4 columns 


the ISP feature of VM would allow you to drill down to individual spindles and 
place subdisks on each spindle.
 
Individual spindles of the RAID group? Doesn't that defeat the purpose of the 
RAID group?
Striping across LUNs gets ...interesting; we usually just use them concat. Of 
course that's with a real SAN array such as Hitachi 99x0 or Sun 61x0.
I'm not sure I see the point of striping LUNs. If you are having performance 
problems from the array, fix the layout of the RAID group on the array: that's 
why you pay the big bucks to Hitachi for their hardware. I not sure I want to 
know about the load that could flatline a RAID-6 array of 6 15K RPM Fiber 
channel disks backed by a multigigabyte RAM cache.
 
I have certainly seen bad storage layout on the host cause hot spots. That's 
when people make ridiculous numbers of small (gigabyte or so) volumes scattered 
all over the place -- another argument against the old way of doing things with 
databases and raw volumes (if you're going to use raw volumes at least use 
decent size ones not 2GB each ). While old ( 10) Solaris AIO did indeed suck 
dead bunnies thorugh a straw for performance, that's no longer a problem in 
Solaris 10 ZFS if you use it natively (using branded zones to run Solaris 8 
and 9 puts the old AIO interface in front) nor would I expect it to be a 
problem with VxFS.
 





From: veritas-vx-boun...@mailman.eng.auburn.edu 
[mailto:veritas-vx-boun...@mailman.eng.auburn.edu] On Behalf Of William Havey
Sent: Wednesday, January 06, 2010 12:00 PM
To: przemol...@poczta.fm
Cc: veritas-vx@mailman.eng.auburn.edu
Subject: Re: [Veritas-vx] Relayout volume from 2 to 4 columns


VM views the two raid groups as single LUNs. It needn't be concerned 
with the layout of each raid group. To change from 2 columns to 4 columns use 
the relayout option to vxassist and also specify the two new LUNs on which to 
place the two new columns.

That being said, the ISP feature of VM would allow you to drill down to 
individual spindles and place subdisks on each spindle.

Bill


On Wed, Jan 6, 2010 at 6:36 AM, przemol...@poczta.fm wrote:


Hello,

we are using VSF 5.0 MP3 on Solaris 10 attached to SAN-based 
hardware array.
On this array we have created 2 raid groups and on each RG we 
have created
a few LUNs:

raid group:   RG1RG2
 LUN1   LUN7
 LUN2   LUN8
 LUN3   LUN9
 LUN4   LUN10
 LUN5   LUN11
 LUN6   LUN12

For performance reason some of our volumes are striped between 
the two raid groups
(using two columns ncol=2) e.g.:

pl name vol ENABLED ACTIVE 419256320 STRIPE 2/128 RW

In this configuration IOs involves two raid groups.

It seems that in the future in certain cases performance might 
be not as expected
so we would like to add two additional LUNs (taken from two 
additional raid groups)
and relayout the whole volume from 2-col to 4-cols e.g.:

raid group:   RG1RG2RG3RG4
 LUN1   LUN7   LUN13  LUN19
 LUN2   LUN8   LUN14  LUN20
 LUN3   LUN9   LUN15  LUN21
 LUN4   LUN10  LUN16  LUN22
 LUN5   LUN11  LUN17  LUN23
 LUN6   LUN12  LUN18  LUN24

Is it possible to order relayout of existings volumes to spread 
it over all four
RGs ? Can I point somehow that it should relayout using these 
particular LUNs ?


Regards
Przemyslaw Bak (przemol)
--
http://przemol.blogspot.com/
 

Re: [Veritas-vx] Relayout volume from 2 to 4 columns

2010-01-06 Thread Joshua Fielden
So, assume your queue depth is 10, and you need a 1TB volume. You can
present a 1TB LUN,  and your whole SCSI queue for that volume is 10. If
you present 2 500GB LUNs, your combined SCSI queue is 20 - one SCSI
mailbox per LUN. If you present 10 100GB LUNs, you get a combined queue
of 100. If you're doing small-block Oracle, with 95-98% cache read rates
on the array (not unheard of), you can process more I/Os in-flight at
any given point in time. You do make the valid point that making changes
such as this can move the bottleneck from being storage-bound to being
cpu-bound, and that's a not-uncommon result of this tuning on
under-provisioned hardware. When I do this exercise with my customers,
we probably run into that problem ~half the time.

 

As to stripe vs. concat specifically -- using a more artificial example,
imagine you're putting together a 100GB volume. If you concatenate it as
2 50GB LUNs, you'll use one of the SCSI queues only, until you've filled
up the first 50GB, then you'll start using both queues (mostly - this is
an example). If you made the volume from, say, 10 10GB LUNs, you have a
greater mathematical chance of any in-flight I/Os sorting across
multiple queues, thus raising your overall I/O the application can
request at any given time.

 

Again, I want to stress that this does not mean this is *always* the
right way to go, YMMV, please consult your local tuning professional or
BOFH, etc. This is just an argument for why you'd ever _want_ to do it,
using artificial examples. Most applications at most entities do not
need this level of storage discipline. Those that do need it often stop
falling over every night, by a simple storage change (or learn they need
more box to meet SLAs, empirically).

 

ZFS masks the gory details, but ZFS == VxFS + VxVM, there are just
different features in each product suite. Underneath the filesystem, it
is still a RAID volume created on the host, that's managed by software -
even if it's just one LUN. ZFS has features SF doesn't, and vice-versa,
and we also have a CLI front-end that removes most of the complexity of
the product for simple setups. My personal opinion is they fill
different niches, ultimately, and neither is better, just better at
some tasks.

 

As to the Oracle issue, that's what the ODM API from Oracle solves,
amongst other solutions . This article
(http://blogs.sun.com/bobs/entry/one_i_o_two_i ) explains the benefits
of concurrent, unbuffered I/O for Oracle better than I am able to.

 

Cheers,

 

jf

 

As always, anything sent to this list are my own thoughts, and I speak
only for myself.

 

From: veritas-vx-boun...@mailman.eng.auburn.edu
[mailto:veritas-vx-boun...@mailman.eng.auburn.edu] On Behalf Of Hudes,
Dana
Sent: Wednesday, January 06, 2010 11:10 AM
To: William Havey
Cc: veritas-vx@mailman.eng.auburn.edu
Subject: Re: [Veritas-vx] Relayout volume from 2 to 4 columns

 

So indeed this feature turns your $$$ Hitachi 9990V into a JBOD.

Wow. I guess there are products that need this sort of thing. 

And enterprises where the people running the SAN array can't manage it
properly so the server administrators have need of bypassing them.

 

The other question of SCSI queues one per column as a benefit of
striping is interesting. 

Doesn't this just generate extra disk i/o? The array is already doing
RAID with 2 parity stripes. Now what?

Yet this is effectively what ZFS does so there must be a performance
gain.  Hmm. Multiple SCSI queues actually might possibly make sense if
you have a large number of CPUs (like the Sun 4v architecture esp. 5240
with 128 cores or 5440 with 256, or a 4+ board domain on the SunFire 25K
which gives you 32 cores) all of which are running threads that do disk
i/o.  

This benefit seems more practicaly in the ZFS approach where you have
one volume-equivalent (the zpool is both disk group and VM volume in
that it has storage layout) and many filesystems so you would likely
have multiple processes doing independent disk i/o.  In the VxVM one
volume one filesstem model your e.g. Oracle table space is in one
filesystem as one huge file (possibly other databases are files in other
filesystems). Even if you have multiple listeners doing their thing
ultimately there's one file they're working on...of course Oracle has
row locking and other paralleization hmm.

 

 

 

 



From: William Havey [mailto:bbha...@gmail.com] 
Sent: Wednesday, January 06, 2010 12:30 PM
To: Hudes, Dana
Cc: veritas-vx@mailman.eng.auburn.edu
Subject: Re: [Veritas-vx] Relayout volume from 2 to 4 columns

Yes, it certainly does. And that is why Symantec put the feature
in the VM product; to use host-based software to construct and control
storage from host to physical disk. This would help eliminate
multi-vendor chaos in the storage aspects of the data center.

On Wed, Jan 6, 2010 at 12:19 PM, Hudes, Dana
hud...@hra.nyc.gov wrote:

 

Re: [Veritas-vx] changing vxdmp iopolicy on live volumes

2008-11-21 Thread Joshua Fielden
Using vxdmpadm disable ctlr=cX to disable one path administratively
(or other individual disable syntax as-needed) is the best-practice for
doing SAN work.

Iopolicy can be set on the fly -- are you sure you have the correct
ASL/APM packages for NetAPP storage for your version? Having the
wrong/missing APM (Array Policy Module) can cause problems during
administrative path work.

Cheers,

jf

-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Romeo
Theriault
Sent: Friday, November 21, 2008 4:11 PM
To: veritas-vx@mailman.eng.auburn.edu
Subject: [Veritas-vx] changing vxdmp iopolicy on live volumes

I was under the impression that changing vxdmp's iopolicy on the fly was
supported. I recently changed a live Solaris 9 hosts vxdmp iopolicy from
round-robin to singleactive and a handful of it's FCP SAN volumes went
offline with read/write errors. I quickly switched the iopolicy back to
round-robin and was able to get the offlined volumes back online with
some fsck's, etc All of the volumes have two paths through different
FCP cards and switches.

So was I wrong, is changing the iopolicy on vxdmp not a safe thing to
do? Should I have done something differently?

This is what I did:
vxdmpadm setattr enclosure FAS30500 iopolicy=singleactive

On a related topic, the person whom I've replaced didn't leave the FCP
switch passwords behind so we need to reboot the switches to their
bootprom to reset the passwords. I was really hoping that I'd be able to
change the iopolicy on the Solaris/Veritas machines to singleactive and
then change their primary controller for a time being so I could reboot
the FCP switches one at a time to reset their passwords, thus removing
the need to shutdown all of the machines. Is this possible?

Thanks for any hints and help on the matter,

-- 
Romeo Theriault
System Administrator
Information Technology Services


___
Veritas-vx maillist  -  Veritas-vx@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx


Re: [Veritas-vx] VxVM Serial Split Brain

2008-07-24 Thread Joshua Fielden
From an internal technote, so no link, sorry:

Background:
The Serial Split Brain condition arises because VERITAS Volume Manager
(tm) increments the serial ID in the disk media record of each imported
disk in all the disk group configurations on those disks. A new serial
(SSB) ID has been included as part of the new disk group version=110 in
Volume Manager 4 to assist with recovery of the disk group from this
condition. The value that is stored in the configuration database
represents the serial ID that the disk group expects a disk to have. The
serial ID that is stored in a disk's private region is considered to be
its actual value. 

If some disks went missing from the disk group (due to physical
disconnection or power failure), the serial IDs for the disks in their
copies of the configuration database, and also in each disk's private
region, are updated separately on that host. When the disks are
subsequently re-imported into the original shared disk group, the actual
serial IDs on the disks do not agree with the expected values from the
configuration copies on other disks in the disk group.

It then goes into the splitlines routine, and the same basic information
that exists in public technotes on the topic.

Cheers,

jf

-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Tim
Bishop
Sent: Thursday, July 24, 2008 7:59 AM
To: veritas-vx@mailman.eng.auburn.edu
Subject: [Veritas-vx] VxVM Serial Split Brain

Hi all,

I'm using VxVM 4.1 MP2 on Solaris 9 and have a strange problem. I'm
trying to add some disks to a diskgroup and I get the following error:

# /usr/lib/vxvm/bin/vxdisksetup -fi c2t216000C0FF892131d10
# /usr/lib/vxvm/bin/vxdisksetup -fi c2t216000C0FF892131d11
# vxdg -g TeachFS-DG adddisk TeachFS-DGa01=c2t216000C0FF892131d10s2
TeachFS-DGa02=c2t216000C0FF892131d11s2
VxVM vxdg ERROR V-5-1-10127 associating disk-media TeachFS-DGa01 with
c2t216000C0FF892131d10s2:
Serial Split Brain detected. Run vxsplitlines

The reason I'm doing this is because we had an array failure which took
out half of our mirror. The disks I'm trying to add back are on the
failed array, but the problem appears to be with the part that didn't
fail (DGa01 and DGa02 are the failed ones, DGb01 and DGb02 the good
ones):

# /usr/lib/vxvm/bin/vxsplitlines -g TeachFS-DG
[ ## ]
  VxVM vxsplitlines NOTICE V-5-2-2708 There are 1 pools.
The Following are the disks in each pool. Each disk in the same pool
has config copies that are similar.
  VxVM vxsplitlines INFO V-5-2-2707 Pool 0.
c2t216000C0FF808C08d10s2 TeachFS-DGb01
To see the configuration copy from this disk issue
/etc/vx/diag.d/vxprivutil dumpconfig
/dev/vx/dmp/c2t216000C0FF808C08d10s2
To import the diskgroup with config copy from this
disk use the following command
/usr/sbin/vxdg -o selectcp=1128803667.95.qetesh import TeachFS-DG
The following are the disks whose ssb ids don't match in this config
copy
TeachFS-DGb01
TeachFS-DGb02

So I checked the ssb ids on the disks themselves:

# vxdisk list c2t216000C0FF808C08d10s2 | grep ssb
ssb:   actual_seqno=0.0
# vxdisk list c2t216000C0FF808C08d11s2 | grep ssb
ssb:   actual_seqno=0.0

And then in the config on the disks:

# /etc/vx/diag.d/vxprivutil dumpconfig
/dev/vx/dmp/c2t216000C0FF808C08d10s2 | grep ssb
  ssbid=0.2
  ssbid=0.2
# /etc/vx/diag.d/vxprivutil dumpconfig
/dev/vx/dmp/c2t216000C0FF808C08d11s2 | grep ssb
  ssbid=0.2
  ssbid=0.2

So the problem looks fairly obvious. I expect deporting and reimporting
will fix it (the configs on the disks are otherwise identical).

But I'm curious how this came about in the first place. This volume has,
as far as I'm aware, not been unmounted during the problems. And I added
disks to the group not that long ago.

So my question is, does anyone have any idea why this might occur?

Thanks,

Tim.

-- 
Tim Bishop
http://www.bishnet.net/tim/
PGP Key: 0x5AE7D984
___
Veritas-vx maillist  -  Veritas-vx@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx

___
Veritas-vx maillist  -  Veritas-vx@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx


Re: [Veritas-vx] Solaris 10 x86 Storage Foundation MP1

2008-07-08 Thread Joshua Fielden
5.0MP1 has not been released for x64 Solaris.

-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Thomas
Glanzmann
Sent: Tuesday, July 08, 2008 9:52 AM
To: veritas-vx@mailman.eng.auburn.edu
Subject: [Veritas-vx] Solaris 10 x86 Storage Foundation MP1

Hello,
is there a Storage Foundation 5MP1 Update available for Solaris 10
_x86_? I downloaded ,,sxrt5.0MP1.dvd1.tar.gz'' which seems to include
the updates for Linux, Windows, HPUX, Aix, Solaris Sparc, but not
Solaris x86. Is there a different distribution for Sparc x86 or is
SF5MP1 not available for Solaris 10 x86?

Thomas
___
Veritas-vx maillist  -  Veritas-vx@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx

___
Veritas-vx maillist  -  Veritas-vx@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx


Re: [Veritas-vx] Upgrading Veritas 5.0

2008-07-08 Thread Joshua Fielden
RPs are cumulative, so only installing the latest RP would cut down some
time, but not a lot.

-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Scott
Kaiser
Sent: Tuesday, July 08, 2008 2:59 PM
To: Chuck Sedlacko; veritas-vx@mailman.eng.auburn.edu
Subject: Re: [Veritas-vx] Upgrading Veritas 5.0

Have you considered Solaris LiveUpgrade and the VxVM LU support script
that came with 5.0? Not a good fit if you have data volumes on an
encapsulated boot disk, but otherwise should simplify.

The installer does have a tendency to overrecommend reboots, I know that
one unnecessary reboot was ID'd a while back (post-5.0GA) perhaps it is
yours but you'd need to check with support. We plan to fix that in a
future release (and do our best to reduce/eliminate upgrade reboots
generally).

Scott

-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Chuck
Sedlacko
Sent: Tuesday, July 08, 2008 6:44 AM
To: veritas-vx@mailman.eng.auburn.edu
Subject: [Veritas-vx] Upgrading Veritas 5.0

I want to upgrade Solaris 9 servers running Storage Foundation 4.0 MP2
(just volume manager and file system) to Veritas 5.0 with the lated MP
and RP patches.

The procedure I've been following below takes about 3 hours and 4
reboots.  Can any of these steps be combined/streamlined to eliminate
1 or more reboots?  

For example, do I really need a reboot between the installer script and
running the installer -configure script ?  What exactly does the
installer -configure script do?  Can I run this script in single user
mode?  And if so, can I add the patches say either immediately after
running the installer or the installer -configure scripts?

Here's the procedure I've been following from  the Veritas Storage
Foundation Installation Guide, pages 83-84
1) Boot the server in single user mode (15 minutes)
2) Mount the installation packages
3) Run the installer script which prompts me to upgrade my storage
foundation sofware. (40 minutes)

After the installer script completes I get this message.

 At least one package will require a reboot prior to configuration. 
 Please reboot and run installer -configure on the following systems 
 after reboot has completed:
 

So I need to do a reboot at this point and run the installer -configure
option? It doesn't say if I need to bring the server all the way up to
multi-user mode or if just single user mode is sufficient.  I know I've
had issues with the GUI until I run this script, so for my initial
upgrades, I brought the server all the way up to multi-user mode.

4) reboot the server to multi-user mode  (15 minutes)
5) remount the installation packages
6) run the installer -configure script  (5-10 minutes)

At this point, I still can't use the GUI (even though all my file
systems are available) until I reboot. I also have to add the latest
patches so I once again reboot to single user mode...

7) reboot the server to single user mode (10 minutes)
8) Mount the patch directory
9) run the installmp script to install MP1 (25 minutes)
10)  Apply the relative patches from RP1, RP2, RP3, RP4 (10 minutes)
11) Reboot the server to multi-user mode (15 minutes).

I'd like to minimize my downtime.  Any help/suggestions/comments are
appreciated.

Thanks,

Chuck

-- 
|Chuck Sedlacko   | Golf is 20 percent mechanics and technique.  
|LEXIS-NEXIS  |  The other 80 percent is philosophy, humor, 
|P.O. Box 933 |  tragedy, romance, melodrama, companionship, 
|Dayton, Ohio 45401   |  camaraderie, cussedness, and conversation.
|937-865-1781 | -- Grantland Rice



___
Veritas-vx maillist  -  Veritas-vx@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx

___
Veritas-vx maillist  -  Veritas-vx@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx

___
Veritas-vx maillist  -  Veritas-vx@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx


Re: [Veritas-vx] VFS snapshots as a way to undo changes?

2008-02-26 Thread Joshua Fielden
Yeah, I inverted the two in my mind when I first sent that. Sorry I
haven't responded with a correction -- travel in the Midwest is
all-consuming this time of year. :)

Cheers,

jf 

-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Scott
Kaiser
Sent: Tuesday, February 26, 2008 12:15 PM
To: A Darren Dunham; Veritas-vx@mailman.eng.auburn.edu
Subject: Re: [Veritas-vx] VFS snapshots as a way to undo changes?

Checkpoints can be mounted r/w, I don't think snapshots can be. 

Try 

# fsckptadm create ckpt1 /data
# mount -F vxfs -orw,ckpt=ckpt1 /dev/vx/dsk/datadg/datavol:ckpt1 /ckpt1

Regards,
Scott

-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of A Darren
Dunham
Sent: Sunday, February 24, 2008 10:12 PM
To: Veritas-vx@mailman.eng.auburn.edu
Subject: Re: [Veritas-vx] VFS snapshots as a way to undo changes?

On Fri, Feb 22, 2008 at 03:13:36PM -0700, Joshua Fielden wrote:
 You can mount snapshots r/w.

Does that require a particular version?

$ mount -F vxfs -o rw,snapof=/mnt /dev/vx/dsk/testdg/test2 /mnt2 vxfs
mount: conflicting suboptions: ro and rw.
vxfs mount: Usage:
[blah...]
$ mount -F vxfs -o snapof=/mnt /dev/vx/dsk/testdg/test2 /mnt2 $ mount |
grep /mnt2
/mnt2 on /dev/vx/dsk/testdg/test2 read
only/setuid/snapof=/mnt/nolargefiles/dev=1b5b969 on Sun Feb 24 22:17:11
2008

-- 
Darren Dunham   [EMAIL PROTECTED]
Senior Technical Consultant TAOShttp://www.taos.com/
Got some Dr Pepper?   San Francisco, CA bay area
  This line left intentionally blank to confuse you. 
___
Veritas-vx maillist  -  Veritas-vx@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx

___
Veritas-vx maillist  -  Veritas-vx@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx

___
Veritas-vx maillist  -  Veritas-vx@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx


Re: [Veritas-vx] You can be lucky

2008-02-26 Thread Joshua Fielden
Also, for the future:

usedisk Profile Keyword
usedisk disk_name ...
 


By default, the JumpStart program uses all of the operational disks on
the system when you specify partitioning default. The usedisk profile
keyword designates one or more disks that you want the JumpStart program
to use. You must specify disk_name in the form cxtydz or cydz, for
example, c0t0d0 or c0d0s0. 

If you specify usedisk in a profile, the JumpStart program uses only the
disks that you specify after the usedisk keyword.

:)

jf
 

-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Hudes,
Dana
Sent: Tuesday, February 26, 2008 1:06 PM
To: Robinson, Greg; veritas-vx@mailman.eng.auburn.edu
Subject: Re: [Veritas-vx] You can be lucky

You really ought to go to version 5 of SF on Solaris 10. Make sure to
install not only the Master Patches but the Rolling Patches as well.
Also make sure to not only run Solaris 10 update 4 but also that you are
using Update Connection to keep your system reasonably up-to-date.

Mirror your root with SVM. Veritas has some claim to advantage for SAN
LUNs over using ZFS, due to the Array Support Libraries that they claim
give extra performance. If you have an application that demands every
ounce of I/O from the SAN and you have a 4 Gbit FC SAN (you mentioned
iSCSI, is that what you're using for your SAN? OMG...hope it's at least
running gigabit speed...the 5120 has the capacity to have 10G Ethernet,
I might do iSCSI on that but not on 100baseT) OR you're doing something
...interesting...like VVR but otherwise? Sure, if you're running Oracle
you've got SFORA.


=
Dana Hudes
UNIX and Imaging group
NYC-HRA MIS
+1 718 510 8586
Nextel:  172*26*16684
=

-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of
Robinson, Greg
Sent: Tuesday, February 26, 2008 2:38 AM
To: veritas-vx@mailman.eng.auburn.edu
Subject: [Veritas-vx] You can be lucky

Hi all,

We had a server crash today (a T2000 - Solaris 10, vxvm 4.1 with some
patches).  The root disk did not survive, and so we had to rebuild the
server.  It wasn't mirrored either - long story involving iSCSI

The server was still zoned to the SAN disks (about 56) of them, and we
jump started it.  Jumpstart selected the root disk, as per our
jumpstart configuration, but the root disk that it chose was a SAN disk!
:-(  Subsequently, one of our plexes was overwritten (Important safety
tip there!):

DEVICE   TYPEDISK GROUPSTATUS
Disk_0   auto:none   --online invalid
EVA3_0   auto:cdsdiskProjects05   Projects online
EVA3_1   auto:cdsdiskProjects04   Projects online
EVA3_2   auto:none   --online invalid
EVA3_3   auto:cdsdiskProjects01   Projects online
EVA3_4   auto:cdsdisksystem01 system   online
EVA3_5   auto:cdsdiskProjects06   Projects online
-- Projects02   Projects failed was:EVA3_0

So, I removed the failed disk and re-initialised it and then replaced
it.

Now, this disk was used by 3 volumes.  2 of them have been recovered but
the 3rd has not:

Disk group: Projects

TY NAME ASSOCKSTATE   LENGTH   PLOFFS   STATETUTIL0
PUTIL0
dg Projects Projects -----
-

dm Projects01   EVA3_3   -524252928 -   --
-
dm Projects02   EVA3_2   -1048540928 -  --
-
dm Projects04   EVA3_1   -1048540928 -  --
-
dm Projects05   EVA3_0   -1048540928 -  --
-
dm Projects06   EVA3_5   -1300166400 -  --
-

v  Model fsgenDISABLED 1048576000 -  ACTIVE   -   -
pl Model-02 Model DISABLED 1048576000 -   RECOVER  -   -
sd Projects02-01 Model-02 ENABLED 209715200 0--   -
sd Projects04-01 Model-02 ENABLED 838860800 209715200 -   -   -

Even though it says RECOVER, a full fsck has failed.  A full fsck on the
other 2 has brought them back to life.

My questions:

1. Can I be confident in the other 2 volumes?  The install of solaris
only lasted for about 45 seconds and only installed about 2 GB.
2. Should I restore from backup all 3 affected volumes just to be sure?
3. Why did a re-initialise of this LUN not lose the data?  Was it only
the header that was initialised and not the data on the rest of the
disk?

Thankx,

Greg.



IMPORTANT: This email remains the property of the Australian Defence
Organisation and is subject to the jurisdiction of section 70 of the
CRIMES ACT 1914.  If you have received this email in error, you are
requested to contact the sender and delete the email.



___
Veritas-vx maillist  -  Veritas-vx@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx

___

Re: [Veritas-vx] [Veritas-ha] problem setting up I/O fencing

2007-07-31 Thread Joshua Fielden
Are they less than 21MB in size?

 -Original Message-
From:   Zhisong Jason Jin [mailto:[EMAIL PROTECTED]
Sent:   Tuesday, July 31, 2007 07:44 AM US Mountain Standard Time
To: [EMAIL PROTECTED]; veritas-vx
Subject:Re: [Veritas-ha] problem setting up I/O fencing

Hi, gurus:

I'm trying to  setup I/O fencing for a newly build cluster
running: Solaris 9, VCS 3.5 GA  and VXVM 3.5, that connected 
to EMC CX700.  we use this older verison to testing upgrade to 4.x.

a few LUNS were carve out from EMC and are seeing on the host.
and I'm try to use the 3 LUNS to setup the coordinator disk.   

 
the three LUN(disk) I need to use are 
c3t5006016110205AADd2 
c3t5006016110205AADd3
c3t5006016110205AADd4

each are showing up in format with 4 paths:

# echo | format | grep ADd2
  14. c2t5006016910205AADd2 DGC-RAID5-0219 cyl 254 alt 2 hd 1 sec
16
  15. c2t5006016010205AADd2 DGC-RAID5-0219 cyl 254 alt 2 hd 1 sec
16
  26. c3t5006016810205AADd2 DGC-RAID5-0219 cyl 254 alt 2 hd 1 sec
16
  27. c3t5006016110205AADd2 DGC-RAID5-0219 cyl 254 alt 2 hd 1 sec
16

however 

# vxdg init vxfencoorddg c3t5006016110205AADd2 c3t5006016110205AADd3
c3t5006016110205AADd4

vxvm:vxdg: ERROR: Device c3t5006016110205AADd2 not found in
configuration
vxvm:vxdg: ERROR: Device c3t5006016110205AADd3 not found in
configuration

# vxdisk list c2t5006016910205AADd4 
vxvm:vxdisk: ERROR: Disk c2t5006016910205AADd4: Disk not in the
configuration

# vxdisk list
DEVICE   TYPE  DISK GROUPSTATUS
EMC_CLARiiON0_0 sliced--error
EMC_CLARiiON0_1 sliced--error
EMC_CLARiiON0_2 sliced--error
c1t0d0s2 sliced--error
c1t0d0s6 simplerootdisk rootdg   online



# vxdisk -e list 
DEVICE   TYPE  DISK GROUPSTATUS
c#t#d#_NAME   
EMC_CLARiiON0_0 slicedEMC_CLARiiON0_0 oraproddgonline
c2t5006016910205AADd5s2
EMC_CLARiiON0_1 slicedEMC_CLARiiON0_1 oraproddgonline
c2t5006016910205AADd1s2
EMC_CLARiiON0_2 slicedEMC_CLARiiON0_2 oraproddgonline
c2t5006016910205AADd0s2
c1t0d0s2 sliced--error
c1t0d0s2
c1t0d0s6 simplerootdisk rootdg   online


what's wrong? your comments or advice is much appreciated. 

Thanks.
Jason

___
Veritas-ha maillist  -  [EMAIL PROTECTED]
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-ha

___
Veritas-vx maillist  -  Veritas-vx@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx


Re: [Veritas-vx] [Veritas-ha] problem setting up I/O fencing

2007-07-31 Thread Joshua Fielden
The RAC docuentation specifies 20MB or greater, but I concede that the standard 
CVM/VCS docs probably don't.
 The interop issue is in 5x73 EMC firmware, the 'gatekeeper bit' was changed to 
mean 'gatekeeper, or high LUN address' as part of supporting LUNs numbered 
higher than 8k (if memory serves, possibly just LUNs higher than 'ff'). There 
is code in some versions of VolMgr to look for that bit set, AND LUN size less 
than (or possibly equal to -- still going from memory here) 20MB, and assume 
the device is a gatekeeper if those two facts intersect.

There is a hotfix, or just grow the LUNs to 21MB.

jf
--
All above statements are my own, and come with no endorsement or warranty, etc.


 -Original Message-
From:   Ceri Davies [mailto:[EMAIL PROTECTED]
Sent:   Tuesday, July 31, 2007 11:22 AM US Mountain Standard Time
To: Joshua Fielden
Cc: Zhisong Jason Jin; [EMAIL PROTECTED]; veritas-vx
Subject:Re: [Veritas-vx] [Veritas-ha] problem setting up I/O fencing

On Tue, Jul 31, 2007 at 09:12:44AM -0700, Joshua Fielden wrote:
 Are they less than 21MB in size?

Is that likely to cause a problem?  The VCS documentation for
I/O fencing specifically says that:

  Symantec recommends using the smallest possible LUNs for coordinator
   disks.

and:

  Any disks that support SCSI-3 Persistent Reservation can be
   coordinator disks.  Best practice is to select the smallest
   possible LUNs for use as coordinator disks.

If that means LUNs over 21MB then it should say that (which is not to
say that this isn't a requirement, just that the VCS documentation
implies that it isn't).

Ceri
-- 
That must be wonderful!  I don't understand it at all.
  -- Moliere

___
Veritas-vx maillist  -  Veritas-vx@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx