Re: [ovirt-users] best way to remove SAN lun
On Sun, Mar 5, 2017 at 2:40 PM, Pavel Gashev wrote: > Please also consider a case when a single iSCSI target has several LUNs and > you remove one of them. > In this case you should not logout. Right. ovirt manages connections to storage. When you remove the last usage of a connection, we should disconnect from the target. If this is not the case, this is ovirt-engine bug. > > -Original Message- > From: on behalf of Nelson Lameiras > > Date: Friday, 3 March 2017 at 20:55 > To: Nir Soffer > Cc: users > Subject: Re: [ovirt-users] best way to remove SAN lun > > Hello Nir, > > I think I was not clear in my explanations, let me try again : > > we have a oVirt 4.0.5.5 cluster with multiple hosts (centos 7.2). > In this cluster, we added a SAN volume (iscsi) a few months ago directly in > GUI > Later we had to remove a DATA volume (SAN iscsi). Below the steps we have > taken : > > 1- we migrated all disks outside the volume (oVirt) > 2- we put volume on maintenance (oVirt) > 3- we detach volume (oVirt) > 4- we removed/destroyed volume (oVirt) > > In SAN : > 5- we put it offline on SAN > 6- we delete it from SAN > > We thought this would be enough, but later we has a serious incident when log > partition went full (partially our fault) : > /var/log/messages was continuously logging that it was still trying to reach > the SAN volumes (we have since taken care of the issue of log space issue => > more aggressive logrotate, etc) > > The real solution was to add two more steps, using shell in ALL hosts : > 4a - logout from SAN : iscsiadm -m node --logout -T iqn. > 4b - remove iscsi targets : rm -fr /var/lib/iscsi/nodes/iqn.X > > This effectively solved our problem, but was fastidious since we had to do it > manually in all hosts (imagine if we had hundreds of hosts...) > > So my question was : shouldn't it be oVirt job to system "logout" and "remove > iscsi targets" automatically when a volume is removed from oVirt ? maybe not, > and I'm missing something? > > cordialement, regards, > > Nelson LAMEIRAS > Ingénieur Systèmes et Réseaux / Systems and Networks engineer > Tel: +33 5 32 09 09 70 > nelson.lamei...@lyra-network.com > > www.lyra-network.com | www.payzen.eu > > > > > > Lyra Network, 109 rue de l'innovation, 31670 Labège, FRANCE > > ----- Original Message - > From: "Nir Soffer" > To: "Nelson Lameiras" > Cc: "Gianluca Cecchi" , "Adam Litke" > , "users" > Sent: Wednesday, February 22, 2017 8:27:26 AM > Subject: Re: [ovirt-users] best way to remove SAN lun > > On Wed, Feb 22, 2017 at 9:03 AM, Nelson Lameiras > wrote: >> Hello, >> >> Not sure it is the same issue, but we have had a "major" issue recently in >> our production system when removing a ISCSI volume from oVirt, and then >> removing it from SAN. > > What version? OS version? > > The order must be: > > 1. remove the LUN from storage domain > will be available in next 4.1 release. in older versions you have > to remove the storage domain > > 2. unzone the LUN on the server > > 3. remove the multipath devices and the paths on the nodes > >> The issue being that each host was still trying to access regularly to the >> SAN volume in spite of not being completely removed from oVirt. > > What do you mean by "not being completely removed"? > > Who was accessing the volume? > >> This led to an massive increase of error logs, which filled completely >> /var/log partition, > > Which log was full with errors? > >> which snowballed into crashing vdsm and other nasty consequences. > > You should have big enough /var/log to avoid such issues. > >> >> Anyway, the solution was to manually logout from SAN (in each host) with >> iscsiadm and manually remove iscsi targets (again in each host). It was not >> difficult once the problem was found because currently we only have 3 hosts >> in this cluster, but I'm wondering what would happen if we had hundreds of >> hosts ? >> >> Maybe I'm being naive but shouldn't this be "oVirt job" ? Is there a RFE >> still waiting to be included on this subject or should I write one ? > > We have RFE for this here: > https://bugzilla.redhat.com/1310330 > > But you must understand that ovirt does not control your storage server, > you are responsible to add devices on the storage server, and remove > them. We are only consuming the devices. > > Even we we provide a way to remove devices on all hosts, you will
Re: [ovirt-users] best way to remove SAN lun
Please also consider a case when a single iSCSI target has several LUNs and you remove one of them. In this case you should not logout. -Original Message- From: on behalf of Nelson Lameiras Date: Friday, 3 March 2017 at 20:55 To: Nir Soffer Cc: users Subject: Re: [ovirt-users] best way to remove SAN lun Hello Nir, I think I was not clear in my explanations, let me try again : we have a oVirt 4.0.5.5 cluster with multiple hosts (centos 7.2). In this cluster, we added a SAN volume (iscsi) a few months ago directly in GUI Later we had to remove a DATA volume (SAN iscsi). Below the steps we have taken : 1- we migrated all disks outside the volume (oVirt) 2- we put volume on maintenance (oVirt) 3- we detach volume (oVirt) 4- we removed/destroyed volume (oVirt) In SAN : 5- we put it offline on SAN 6- we delete it from SAN We thought this would be enough, but later we has a serious incident when log partition went full (partially our fault) : /var/log/messages was continuously logging that it was still trying to reach the SAN volumes (we have since taken care of the issue of log space issue => more aggressive logrotate, etc) The real solution was to add two more steps, using shell in ALL hosts : 4a - logout from SAN : iscsiadm -m node --logout -T iqn. 4b - remove iscsi targets : rm -fr /var/lib/iscsi/nodes/iqn.X This effectively solved our problem, but was fastidious since we had to do it manually in all hosts (imagine if we had hundreds of hosts...) So my question was : shouldn't it be oVirt job to system "logout" and "remove iscsi targets" automatically when a volume is removed from oVirt ? maybe not, and I'm missing something? cordialement, regards, Nelson LAMEIRAS Ingénieur Systèmes et Réseaux / Systems and Networks engineer Tel: +33 5 32 09 09 70 nelson.lamei...@lyra-network.com www.lyra-network.com | www.payzen.eu Lyra Network, 109 rue de l'innovation, 31670 Labège, FRANCE - Original Message - From: "Nir Soffer" To: "Nelson Lameiras" Cc: "Gianluca Cecchi" , "Adam Litke" , "users" Sent: Wednesday, February 22, 2017 8:27:26 AM Subject: Re: [ovirt-users] best way to remove SAN lun On Wed, Feb 22, 2017 at 9:03 AM, Nelson Lameiras wrote: > Hello, > > Not sure it is the same issue, but we have had a "major" issue recently in > our production system when removing a ISCSI volume from oVirt, and then > removing it from SAN. What version? OS version? The order must be: 1. remove the LUN from storage domain will be available in next 4.1 release. in older versions you have to remove the storage domain 2. unzone the LUN on the server 3. remove the multipath devices and the paths on the nodes > The issue being that each host was still trying to access regularly to the > SAN volume in spite of not being completely removed from oVirt. What do you mean by "not being completely removed"? Who was accessing the volume? > This led to an massive increase of error logs, which filled completely > /var/log partition, Which log was full with errors? > which snowballed into crashing vdsm and other nasty consequences. You should have big enough /var/log to avoid such issues. > > Anyway, the solution was to manually logout from SAN (in each host) with > iscsiadm and manually remove iscsi targets (again in each host). It was not > difficult once the problem was found because currently we only have 3 hosts > in this cluster, but I'm wondering what would happen if we had hundreds of > hosts ? > > Maybe I'm being naive but shouldn't this be "oVirt job" ? Is there a RFE > still waiting to be included on this subject or should I write one ? We have RFE for this here: https://bugzilla.redhat.com/1310330 But you must understand that ovirt does not control your storage server, you are responsible to add devices on the storage server, and remove them. We are only consuming the devices. Even we we provide a way to remove devices on all hosts, you will have to remove the device on the storage server before removing it from hosts. If not, ovirt will find the removed devices again in the next scsi rescan, and we do lot of these to support automatic discovery of new devices or resized devices. Nir > > cordialement, regards, > > > Nelson LAMEIRAS > Ingénieur Systèmes et Réseaux / Systems and Networks engineer > Tel: +33 5 32 09 09 70 > nelson.lamei...@lyra-network.com > > www.lyra-network.com | www.payzen.eu > > > > > > Lyra Network, 109 rue de l'innovation, 31670 Labège, FRANCE > > - Original Message - > From: "Nir Soffer" > To: "Gianluca Cecchi" , "Adam Litke" > > Cc: "users" > Sent: Tuesday, February 21, 2
Re: [ovirt-users] best way to remove SAN lun
Hello Nir, I think I was not clear in my explanations, let me try again : we have a oVirt 4.0.5.5 cluster with multiple hosts (centos 7.2). In this cluster, we added a SAN volume (iscsi) a few months ago directly in GUI Later we had to remove a DATA volume (SAN iscsi). Below the steps we have taken : 1- we migrated all disks outside the volume (oVirt) 2- we put volume on maintenance (oVirt) 3- we detach volume (oVirt) 4- we removed/destroyed volume (oVirt) In SAN : 5- we put it offline on SAN 6- we delete it from SAN We thought this would be enough, but later we has a serious incident when log partition went full (partially our fault) : /var/log/messages was continuously logging that it was still trying to reach the SAN volumes (we have since taken care of the issue of log space issue => more aggressive logrotate, etc) The real solution was to add two more steps, using shell in ALL hosts : 4a - logout from SAN : iscsiadm -m node --logout -T iqn. 4b - remove iscsi targets : rm -fr /var/lib/iscsi/nodes/iqn.X This effectively solved our problem, but was fastidious since we had to do it manually in all hosts (imagine if we had hundreds of hosts...) So my question was : shouldn't it be oVirt job to system "logout" and "remove iscsi targets" automatically when a volume is removed from oVirt ? maybe not, and I'm missing something? cordialement, regards, Nelson LAMEIRAS Ingénieur Systèmes et Réseaux / Systems and Networks engineer Tel: +33 5 32 09 09 70 nelson.lamei...@lyra-network.com www.lyra-network.com | www.payzen.eu Lyra Network, 109 rue de l'innovation, 31670 Labège, FRANCE - Original Message - From: "Nir Soffer" To: "Nelson Lameiras" Cc: "Gianluca Cecchi" , "Adam Litke" , "users" Sent: Wednesday, February 22, 2017 8:27:26 AM Subject: Re: [ovirt-users] best way to remove SAN lun On Wed, Feb 22, 2017 at 9:03 AM, Nelson Lameiras wrote: > Hello, > > Not sure it is the same issue, but we have had a "major" issue recently in > our production system when removing a ISCSI volume from oVirt, and then > removing it from SAN. What version? OS version? The order must be: 1. remove the LUN from storage domain will be available in next 4.1 release. in older versions you have to remove the storage domain 2. unzone the LUN on the server 3. remove the multipath devices and the paths on the nodes > The issue being that each host was still trying to access regularly to the > SAN volume in spite of not being completely removed from oVirt. What do you mean by "not being completely removed"? Who was accessing the volume? > This led to an massive increase of error logs, which filled completely > /var/log partition, Which log was full with errors? > which snowballed into crashing vdsm and other nasty consequences. You should have big enough /var/log to avoid such issues. > > Anyway, the solution was to manually logout from SAN (in each host) with > iscsiadm and manually remove iscsi targets (again in each host). It was not > difficult once the problem was found because currently we only have 3 hosts > in this cluster, but I'm wondering what would happen if we had hundreds of > hosts ? > > Maybe I'm being naive but shouldn't this be "oVirt job" ? Is there a RFE > still waiting to be included on this subject or should I write one ? We have RFE for this here: https://bugzilla.redhat.com/1310330 But you must understand that ovirt does not control your storage server, you are responsible to add devices on the storage server, and remove them. We are only consuming the devices. Even we we provide a way to remove devices on all hosts, you will have to remove the device on the storage server before removing it from hosts. If not, ovirt will find the removed devices again in the next scsi rescan, and we do lot of these to support automatic discovery of new devices or resized devices. Nir > > cordialement, regards, > > > Nelson LAMEIRAS > Ingénieur Systèmes et Réseaux / Systems and Networks engineer > Tel: +33 5 32 09 09 70 > nelson.lamei...@lyra-network.com > > www.lyra-network.com | www.payzen.eu > > > > > > Lyra Network, 109 rue de l'innovation, 31670 Labège, FRANCE > > - Original Message - > From: "Nir Soffer" > To: "Gianluca Cecchi" , "Adam Litke" > > Cc: "users" > Sent: Tuesday, February 21, 2017 6:32:18 PM > Subject: Re: [ovirt-users] best way to remove SAN lun > > On Tue, Feb 21, 2017 at 7:25 PM, Gianluca Cecchi > wrote: >> On Tue, Feb 21, 2017 at 6:10 PM, Nir Soffer wrote: >>> >>> This is caused by active lvs on the remove storage domains that were not >>> de
Re: [ovirt-users] best way to remove SAN lun
On Wed, Feb 22, 2017 at 9:27 AM Nir Soffer wrote: > On Wed, Feb 22, 2017 at 9:03 AM, Nelson Lameiras > wrote: > > Hello, > > > > Not sure it is the same issue, but we have had a "major" issue recently > in our production system when removing a ISCSI volume from oVirt, and then > removing it from SAN. > > What version? OS version? > > The order must be: > > 1. remove the LUN from storage domain > will be available in next 4.1 release. in older versions you have > to remove the storage domain > > 2. unzone the LUN on the server > > 3. remove the multipath devices and the paths on the nodes > > > The issue being that each host was still trying to access regularly to > the SAN volume in spite of not being completely removed from oVirt. > > What do you mean by "not being completely removed"? > > Who was accessing the volume? > > > This led to an massive increase of error logs, which filled completely > /var/log partition, > > Which log was full with errors? > > > which snowballed into crashing vdsm and other nasty consequences. > > You should have big enough /var/log to avoid such issues. > - Log rotation should be set better not to consume excessive amounts of space. I'm seeing /etc/vdsm/logrotate/vdsm - not sure why it's not under /etc/logrotate.d . Looking at the file, seems like there's a 15M limit and 100 files, which translates to 1.5GB - and it is supposed to be compressed (not sure XZ is a good choice - it's very CPU intensive). Others (Gluster?) do not seem to have a size limit, just weekly. Need to look at other components as well. - At least on ovirt-node, we'd like to separate some directories to different partitions. So for example core dumps (which should be limited as well) on /var/core do not fill the same partition as /var/log is and thus render the host unusable. And again, looking at file, we have a 'size 0' on /var/log/core/*.dump - and 'rotate 1' - not sure what it means - but it should not be in /var/log/core, but /var/core, I reckon. Y. > > > > Anyway, the solution was to manually logout from SAN (in each host) with > iscsiadm and manually remove iscsi targets (again in each host). It was not > difficult once the problem was found because currently we only have 3 hosts > in this cluster, but I'm wondering what would happen if we had hundreds of > hosts ? > > > > Maybe I'm being naive but shouldn't this be "oVirt job" ? Is there a RFE > still waiting to be included on this subject or should I write one ? > > We have RFE for this here: > https://bugzilla.redhat.com/1310330 > > But you must understand that ovirt does not control your storage server, > you are responsible to add devices on the storage server, and remove > them. We are only consuming the devices. > > Even we we provide a way to remove devices on all hosts, you will have > to remove the device on the storage server before removing it from > hosts. If not, ovirt will find the removed devices again in the next > scsi rescan, > and we do lot of these to support automatic discovery of new devices > or resized devices. > > Nir > > > > > cordialement, regards, > > > > > > Nelson LAMEIRAS > > Ingénieur Systèmes et Réseaux / Systems and Networks engineer > > Tel: +33 5 32 09 09 70 <+33%205%2032%2009%2009%2070> > > nelson.lamei...@lyra-network.com > > > > www.lyra-network.com | www.payzen.eu > > > > > > > > > > > > Lyra Network, 109 rue de l'innovation, 31670 Labège, FRANCE > > > > - Original Message - > > From: "Nir Soffer" > > To: "Gianluca Cecchi" , "Adam Litke" < > ali...@redhat.com> > > Cc: "users" > > Sent: Tuesday, February 21, 2017 6:32:18 PM > > Subject: Re: [ovirt-users] best way to remove SAN lun > > > > On Tue, Feb 21, 2017 at 7:25 PM, Gianluca Cecchi > > wrote: > >> On Tue, Feb 21, 2017 at 6:10 PM, Nir Soffer wrote: > >>> > >>> This is caused by active lvs on the remove storage domains that were > not > >>> deactivated during the removal. This is a very old known issue. > >>> > >>> You have remove the remove device mapper entries - you can see the > devices > >>> using: > >>> > >>> dmsetup status > >>> > >>> Then you can remove the mapping using: > >>> > >>> dmsetup remove device-name > >>> > >>> Once you removed the stale lvs, you will be able to remove the > multipath > >>> dev
Re: [ovirt-users] best way to remove SAN lun
On Wed, Feb 22, 2017 at 9:03 AM, Nelson Lameiras wrote: > Hello, > > Not sure it is the same issue, but we have had a "major" issue recently in > our production system when removing a ISCSI volume from oVirt, and then > removing it from SAN. What version? OS version? The order must be: 1. remove the LUN from storage domain will be available in next 4.1 release. in older versions you have to remove the storage domain 2. unzone the LUN on the server 3. remove the multipath devices and the paths on the nodes > The issue being that each host was still trying to access regularly to the > SAN volume in spite of not being completely removed from oVirt. What do you mean by "not being completely removed"? Who was accessing the volume? > This led to an massive increase of error logs, which filled completely > /var/log partition, Which log was full with errors? > which snowballed into crashing vdsm and other nasty consequences. You should have big enough /var/log to avoid such issues. > > Anyway, the solution was to manually logout from SAN (in each host) with > iscsiadm and manually remove iscsi targets (again in each host). It was not > difficult once the problem was found because currently we only have 3 hosts > in this cluster, but I'm wondering what would happen if we had hundreds of > hosts ? > > Maybe I'm being naive but shouldn't this be "oVirt job" ? Is there a RFE > still waiting to be included on this subject or should I write one ? We have RFE for this here: https://bugzilla.redhat.com/1310330 But you must understand that ovirt does not control your storage server, you are responsible to add devices on the storage server, and remove them. We are only consuming the devices. Even we we provide a way to remove devices on all hosts, you will have to remove the device on the storage server before removing it from hosts. If not, ovirt will find the removed devices again in the next scsi rescan, and we do lot of these to support automatic discovery of new devices or resized devices. Nir > > cordialement, regards, > > > Nelson LAMEIRAS > Ingénieur Systèmes et Réseaux / Systems and Networks engineer > Tel: +33 5 32 09 09 70 > nelson.lamei...@lyra-network.com > > www.lyra-network.com | www.payzen.eu > > > > > > Lyra Network, 109 rue de l'innovation, 31670 Labège, FRANCE > > - Original Message - > From: "Nir Soffer" > To: "Gianluca Cecchi" , "Adam Litke" > > Cc: "users" > Sent: Tuesday, February 21, 2017 6:32:18 PM > Subject: Re: [ovirt-users] best way to remove SAN lun > > On Tue, Feb 21, 2017 at 7:25 PM, Gianluca Cecchi > wrote: >> On Tue, Feb 21, 2017 at 6:10 PM, Nir Soffer wrote: >>> >>> This is caused by active lvs on the remove storage domains that were not >>> deactivated during the removal. This is a very old known issue. >>> >>> You have remove the remove device mapper entries - you can see the devices >>> using: >>> >>> dmsetup status >>> >>> Then you can remove the mapping using: >>> >>> dmsetup remove device-name >>> >>> Once you removed the stale lvs, you will be able to remove the multipath >>> device and the underlying paths, and lvm will not complain about read >>> errors. >>> >>> Nir >> >> >> OK Nir, thanks for advising. >> >> So what I run with success on the 2 hosts >> >> [root@ovmsrv05 vdsm]# for dev in $(dmsetup status | grep >> 900b1853--e192--4661--a0f9--7c7c396f6f49 | cut -d ":" -f 1) >> do >>dmsetup remove $dev >> done >> [root@ovmsrv05 vdsm]# >> >> and now I can run >> >> [root@ovmsrv05 vdsm]# multipath -f 3600a0b8000299902cd3c5501458f >> [root@ovmsrv05 vdsm]# >> >> Also, with related names depending on host, >> >> previous maps to single devices were for example in ovmsrv05: >> >> 3600a0b8000299902cd3c5501458f dm-4 IBM ,1814 FAStT >> size=2.0T features='2 pg_init_retries 50' hwhandler='1 rdac' wp=rw >> |-+- policy='service-time 0' prio=0 status=enabled >> | |- 0:0:0:2 sdb8:16 failed undef running >> | `- 1:0:0:2 sdh8:112 failed undef running >> `-+- policy='service-time 0' prio=0 status=enabled >> |- 0:0:1:2 sdg8:96 failed undef running >> `- 1:0:1:2 sdn8:208 failed undef running >> >> And removal of single path devices: >> >> [root@ovmsrv05 root]# for dev in sdb sdh sdg sdn >> do >> echo 1 > /sys/block/${dev}/device/delete >> done >> [root@ovmsrv05 vdsm]# >> >> All clean now... ;-) > > Great! > > I think we should have a script doing all these steps. > > Nir > ___ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] best way to remove SAN lun
Hello, Not sure it is the same issue, but we have had a "major" issue recently in our production system when removing a ISCSI volume from oVirt, and then removing it from SAN. The issue being that each host was still trying to access regularly to the SAN volume in spite of not being completely removed from oVirt. This led to an massive increase of error logs, which filled completely /var/log partition, which snowballed into crashing vdsm and other nasty consequences. Anyway, the solution was to manually logout from SAN (in each host) with iscsiadm and manually remove iscsi targets (again in each host). It was not difficult once the problem was found because currently we only have 3 hosts in this cluster, but I'm wondering what would happen if we had hundreds of hosts ? Maybe I'm being naive but shouldn't this be "oVirt job" ? Is there a RFE still waiting to be included on this subject or should I write one ? cordialement, regards, Nelson LAMEIRAS Ingénieur Systèmes et Réseaux / Systems and Networks engineer Tel: +33 5 32 09 09 70 nelson.lamei...@lyra-network.com www.lyra-network.com | www.payzen.eu Lyra Network, 109 rue de l'innovation, 31670 Labège, FRANCE - Original Message - From: "Nir Soffer" To: "Gianluca Cecchi" , "Adam Litke" Cc: "users" Sent: Tuesday, February 21, 2017 6:32:18 PM Subject: Re: [ovirt-users] best way to remove SAN lun On Tue, Feb 21, 2017 at 7:25 PM, Gianluca Cecchi wrote: > On Tue, Feb 21, 2017 at 6:10 PM, Nir Soffer wrote: >> >> This is caused by active lvs on the remove storage domains that were not >> deactivated during the removal. This is a very old known issue. >> >> You have remove the remove device mapper entries - you can see the devices >> using: >> >> dmsetup status >> >> Then you can remove the mapping using: >> >> dmsetup remove device-name >> >> Once you removed the stale lvs, you will be able to remove the multipath >> device and the underlying paths, and lvm will not complain about read >> errors. >> >> Nir > > > OK Nir, thanks for advising. > > So what I run with success on the 2 hosts > > [root@ovmsrv05 vdsm]# for dev in $(dmsetup status | grep > 900b1853--e192--4661--a0f9--7c7c396f6f49 | cut -d ":" -f 1) > do >dmsetup remove $dev > done > [root@ovmsrv05 vdsm]# > > and now I can run > > [root@ovmsrv05 vdsm]# multipath -f 3600a0b8000299902cd3c5501458f > [root@ovmsrv05 vdsm]# > > Also, with related names depending on host, > > previous maps to single devices were for example in ovmsrv05: > > 3600a0b8000299902cd3c5501458f dm-4 IBM ,1814 FAStT > size=2.0T features='2 pg_init_retries 50' hwhandler='1 rdac' wp=rw > |-+- policy='service-time 0' prio=0 status=enabled > | |- 0:0:0:2 sdb8:16 failed undef running > | `- 1:0:0:2 sdh8:112 failed undef running > `-+- policy='service-time 0' prio=0 status=enabled > |- 0:0:1:2 sdg8:96 failed undef running > `- 1:0:1:2 sdn8:208 failed undef running > > And removal of single path devices: > > [root@ovmsrv05 root]# for dev in sdb sdh sdg sdn > do > echo 1 > /sys/block/${dev}/device/delete > done > [root@ovmsrv05 vdsm]# > > All clean now... ;-) Great! I think we should have a script doing all these steps. Nir ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] best way to remove SAN lun
On Tue, Feb 21, 2017 at 7:25 PM, Gianluca Cecchi wrote: > On Tue, Feb 21, 2017 at 6:10 PM, Nir Soffer wrote: >> >> This is caused by active lvs on the remove storage domains that were not >> deactivated during the removal. This is a very old known issue. >> >> You have remove the remove device mapper entries - you can see the devices >> using: >> >> dmsetup status >> >> Then you can remove the mapping using: >> >> dmsetup remove device-name >> >> Once you removed the stale lvs, you will be able to remove the multipath >> device and the underlying paths, and lvm will not complain about read >> errors. >> >> Nir > > > OK Nir, thanks for advising. > > So what I run with success on the 2 hosts > > [root@ovmsrv05 vdsm]# for dev in $(dmsetup status | grep > 900b1853--e192--4661--a0f9--7c7c396f6f49 | cut -d ":" -f 1) > do >dmsetup remove $dev > done > [root@ovmsrv05 vdsm]# > > and now I can run > > [root@ovmsrv05 vdsm]# multipath -f 3600a0b8000299902cd3c5501458f > [root@ovmsrv05 vdsm]# > > Also, with related names depending on host, > > previous maps to single devices were for example in ovmsrv05: > > 3600a0b8000299902cd3c5501458f dm-4 IBM ,1814 FAStT > size=2.0T features='2 pg_init_retries 50' hwhandler='1 rdac' wp=rw > |-+- policy='service-time 0' prio=0 status=enabled > | |- 0:0:0:2 sdb8:16 failed undef running > | `- 1:0:0:2 sdh8:112 failed undef running > `-+- policy='service-time 0' prio=0 status=enabled > |- 0:0:1:2 sdg8:96 failed undef running > `- 1:0:1:2 sdn8:208 failed undef running > > And removal of single path devices: > > [root@ovmsrv05 root]# for dev in sdb sdh sdg sdn > do > echo 1 > /sys/block/${dev}/device/delete > done > [root@ovmsrv05 vdsm]# > > All clean now... ;-) Great! I think we should have a script doing all these steps. Nir ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] best way to remove SAN lun
On Tue, Feb 21, 2017 at 6:10 PM, Nir Soffer wrote: > This is caused by active lvs on the remove storage domains that were not > deactivated during the removal. This is a very old known issue. > > You have remove the remove device mapper entries - you can see the devices > using: > > dmsetup status > > Then you can remove the mapping using: > > dmsetup remove device-name > > Once you removed the stale lvs, you will be able to remove the multipath > device and the underlying paths, and lvm will not complain about read > errors. > > Nir > OK Nir, thanks for advising. So what I run with success on the 2 hosts [root@ovmsrv05 vdsm]# for dev in $(dmsetup status | grep 900b1853--e192--4661--a0f9--7c7c396f6f49 | cut -d ":" -f 1) do dmsetup remove $dev done [root@ovmsrv05 vdsm]# and now I can run [root@ovmsrv05 vdsm]# multipath -f 3600a0b8000299902cd3c5501458f [root@ovmsrv05 vdsm]# Also, with related names depending on host, previous maps to single devices were for example in ovmsrv05: 3600a0b8000299902cd3c5501458f dm-4 IBM ,1814 FAStT size=2.0T features='2 pg_init_retries 50' hwhandler='1 rdac' wp=rw |-+- policy='service-time 0' prio=0 status=enabled | |- 0:0:0:2 sdb8:16 failed undef running | `- 1:0:0:2 sdh8:112 failed undef running `-+- policy='service-time 0' prio=0 status=enabled |- 0:0:1:2 sdg8:96 failed undef running `- 1:0:1:2 sdn8:208 failed undef running And removal of single path devices: [root@ovmsrv05 root]# for dev in sdb sdh sdg sdn do echo 1 > /sys/block/${dev}/device/delete done [root@ovmsrv05 vdsm]# All clean now... ;-) Thanks again, Gianluca ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] best way to remove SAN lun
On Tue, Feb 21, 2017 at 7:06 PM, Gianluca Cecchi wrote: > The id of the removed storage domain was > 900b1853-e192-4661-a0f9-7c7c396f6f49 and on the not yet reboted hosts I get > this > The multipath device was 3600a0b8000299902cd3c5501458f > > With pvs command I get this errors related to them: > > [root@ovmsrv06 ~]# pvs > /dev/mapper/3600a0b8000299902cd3c5501458f: read failed after 0 of 4096 > at 0: Input/output error > /dev/mapper/3600a0b8000299902cd3c5501458f: read failed after 0 of 4096 > at 2199023190016: Input/output error > /dev/mapper/3600a0b8000299902cd3c5501458f: read failed after 0 of 4096 > at 2199023247360: Input/output error > /dev/mapper/3600a0b8000299902cd3c5501458f: read failed after 0 of 4096 > at 4096: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/metadata: read failed after 0 of > 4096 at 0: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/metadata: read failed after 0 of > 4096 at 536805376: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/metadata: read failed after 0 of > 4096 at 536862720: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/metadata: read failed after 0 of > 4096 at 4096: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/outbox: read failed after 0 of > 4096 at 0: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/outbox: read failed after 0 of > 4096 at 134152192: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/outbox: read failed after 0 of > 4096 at 134209536: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/outbox: read failed after 0 of > 4096 at 4096: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/xleases: read failed after 0 of > 4096 at 0: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/xleases: read failed after 0 of > 4096 at 1073676288: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/xleases: read failed after 0 of > 4096 at 1073733632: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/xleases: read failed after 0 of > 4096 at 4096: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/leases: read failed after 0 of > 4096 at 0: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/leases: read failed after 0 of > 4096 at 2147418112: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/leases: read failed after 0 of > 4096 at 2147475456: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/leases: read failed after 0 of > 4096 at 4096: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/ids: read failed after 0 of 4096 > at 0: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/ids: read failed after 0 of 4096 > at 134152192: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/ids: read failed after 0 of 4096 > at 134209536: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/ids: read failed after 0 of 4096 > at 4096: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/inbox: read failed after 0 of > 4096 at 0: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/inbox: read failed after 0 of > 4096 at 134152192: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/inbox: read failed after 0 of > 4096 at 134209536: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/inbox: read failed after 0 of > 4096 at 4096: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/master: read failed after 0 of > 4096 at 0: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/master: read failed after 0 of > 4096 at 1073676288: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/master: read failed after 0 of > 4096 at 1073733632: Input/output error > /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/master: read failed after 0 of > 4096 at 4096: Input/output error > PVVG > Fmt Attr PSize PFree > /dev/cciss/c0d0p2 cl > lvm2 a-- 67.33g0 > /dev/mapper/3600a0b8000299aa8d08b55014119 > 922b5269-ab56-4c4d-838f-49d33427e2ab lvm2 a-- 4.00t 3.49t > [root@ovmsrv06 ~]# > > How to clean? This is caused by active lvs on the remove storage domains that were not deactivated during the removal. This is a very old known issue. You have remove the remove device mapper entries - you can see the devices using: dmsetup status Then you can remove the mapping using: dmsetup remove device-name Once you removed the stale lvs, you will be able to remove the multipath device and the underlying paths, and lvm will not complain about read errors. Nir > > Gianluca > > > ___ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users > ___ Users mailing list Users@ovirt.org http://lists.ovi
Re: [ovirt-users] best way to remove SAN lun
The id of the removed storage domain was 900b1853-e192-4661-a0f9-7c7c396f6f49 and on the not yet reboted hosts I get this The multipath device was 3600a0b8000299902cd3c5501458f With pvs command I get this errors related to them: [root@ovmsrv06 ~]# pvs /dev/mapper/3600a0b8000299902cd3c5501458f: read failed after 0 of 4096 at 0: Input/output error /dev/mapper/3600a0b8000299902cd3c5501458f: read failed after 0 of 4096 at 2199023190016: Input/output error /dev/mapper/3600a0b8000299902cd3c5501458f: read failed after 0 of 4096 at 2199023247360: Input/output error /dev/mapper/3600a0b8000299902cd3c5501458f: read failed after 0 of 4096 at 4096: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/metadata: read failed after 0 of 4096 at 0: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/metadata: read failed after 0 of 4096 at 536805376: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/metadata: read failed after 0 of 4096 at 536862720: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/metadata: read failed after 0 of 4096 at 4096: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/outbox: read failed after 0 of 4096 at 0: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/outbox: read failed after 0 of 4096 at 134152192: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/outbox: read failed after 0 of 4096 at 134209536: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/outbox: read failed after 0 of 4096 at 4096: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/xleases: read failed after 0 of 4096 at 0: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/xleases: read failed after 0 of 4096 at 1073676288: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/xleases: read failed after 0 of 4096 at 1073733632: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/xleases: read failed after 0 of 4096 at 4096: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/leases: read failed after 0 of 4096 at 0: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/leases: read failed after 0 of 4096 at 2147418112: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/leases: read failed after 0 of 4096 at 2147475456: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/leases: read failed after 0 of 4096 at 4096: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/ids: read failed after 0 of 4096 at 0: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/ids: read failed after 0 of 4096 at 134152192: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/ids: read failed after 0 of 4096 at 134209536: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/ids: read failed after 0 of 4096 at 4096: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/inbox: read failed after 0 of 4096 at 0: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/inbox: read failed after 0 of 4096 at 134152192: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/inbox: read failed after 0 of 4096 at 134209536: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/inbox: read failed after 0 of 4096 at 4096: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/master: read failed after 0 of 4096 at 0: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/master: read failed after 0 of 4096 at 1073676288: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/master: read failed after 0 of 4096 at 1073733632: Input/output error /dev/900b1853-e192-4661-a0f9-7c7c396f6f49/master: read failed after 0 of 4096 at 4096: Input/output error PVVG Fmt Attr PSize PFree /dev/cciss/c0d0p2 cl lvm2 a-- 67.33g0 /dev/mapper/3600a0b8000299aa8d08b55014119 922b5269-ab56-4c4d-838f-49d33427e2ab lvm2 a-- 4.00t 3.49t [root@ovmsrv06 ~]# How to clean? Gianluca ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] best way to remove SAN lun
On Tue, Feb 21, 2017 at 5:42 PM, Andrea Ghelardi wrote: > Hello Gianluca, > > I have ISCSI SAN here and not FC but process should be very similar. > > Correct operation is > > Storage maintenance -> detach -> remove > > > > Once you have removed the storage, Ovirt will release the sanlock. > > Note that your server will still “see” the device and multipath will still > list it as active. > > > > You can now unmap LUN using your SAN manager. > > If your multipath is correctly configured, device will “fail” but nothing > else (multipath and server won’t hang). > > You will be able to perform basic list operations via server shell without > experiencing freeze or locks.* > Yes, but /var/log/messages fills up with multipath errors. This is not desirable > Ovirt will continue to scan new devices so you won’t be able to manually > remove them until you unmap devices from SAN manager. > > > > If you need to remove devices manually, it is advised to follow this guide. > > https://access.redhat.com/documentation/en-US/Red_Hat_ > Enterprise_Linux/6/html/Storage_Administration_Guide/removing_devices.html > > > > * If you do face command freeze etc, re-enable LUN mappings and check your > multipath.conf > Yes, but the problem is that I receive error for the LUN previously configured as storage domain. I presume sum stale devices of kind "dm-??" have remained somewhere. I could reboot one ot the 3 hosts and verified that now the layout is correct on it. But I can't easily reboot just now the remaining 2 hosts, so it would be better to find what forces the multipath device in use Gianluca ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] best way to remove SAN lun
Hello Gianluca, I have ISCSI SAN here and not FC but process should be very similar. Correct operation is Storage maintenance -> detach -> remove Once you have removed the storage, Ovirt will release the sanlock. Note that your server will still “see” the device and multipath will still list it as active. You can now unmap LUN using your SAN manager. If your multipath is correctly configured, device will “fail” but nothing else (multipath and server won’t hang). You will be able to perform basic list operations via server shell without experiencing freeze or locks.* Ovirt will continue to scan new devices so you won’t be able to manually remove them until you unmap devices from SAN manager. If you need to remove devices manually, it is advised to follow this guide. https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Linux/6/html/Storage_Administration_Guide/removing_devices.html * If you do face command freeze etc, re-enable LUN mappings and check your multipath.conf Cheers AG From: users-boun...@ovirt.org [mailto:users-boun...@ovirt.org] On Behalf Of Gianluca Cecchi Sent: Tuesday, February 21, 2017 4:19 PM To: users Subject: [ovirt-users] best way to remove SAN lun Hello, currently I have a cluster of 3 hosts where each one has FC SAN connectivity to 4 LUNs: 3 are already configured as storage domains (1TB, 2TB, 4TB), one is free, not allocated. See here for screenshot: https://drive.google.com/file/d/0BwoPbcrMv8mvRVZZMTlNcTQ5MGs/view?usp=sharing At the moment the command "multipath -l" run on hosts shows all the 4 LUNs. Now I want to do 2 things at storage array level: - remove the 2TB storage domain LUN - remove the 20Gb LUN not yet allocated What is the correct workflow, supposing I have already emptied the 2TB from VM disks ad such? Select 2Tb SD, then Datacenter subtab, then "maintenance", detach" and at the end "remove"? I think I continue to see 4 LUNs at this point, correct? Now I proceed with removal of lun at storage array level? Should I select an SD line and then "Scan disks" to see refresh the SAN and see in multipath only 2 of them at the end? Or any manual command at host level before removal from array? Thanks in advance Gianluca ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] best way to remove SAN lun
On Tue, Feb 21, 2017 at 6:12 PM, Adam Litke wrote: > > > On Tue, Feb 21, 2017 at 10:19 AM, Gianluca Cecchi > wrote: >> >> Hello, >> currently I have a cluster of 3 hosts where each one has FC SAN >> connectivity to 4 LUNs: 3 are already configured as storage domains (1TB, >> 2TB, 4TB), one is free, not allocated. >> See here for screenshot: >> >> https://drive.google.com/file/d/0BwoPbcrMv8mvRVZZMTlNcTQ5MGs/view?usp=sharing >> >> At the moment the command "multipath -l" run on hosts shows all the 4 >> LUNs. >> >> Now I want to do 2 things at storage array level: >> >> - remove the 2TB storage domain LUN >> - remove the 20Gb LUN not yet allocated >> >> What is the correct workflow, supposing I have already emptied the 2TB >> from VM disks ad such? >> Select 2Tb SD, then Datacenter subtab, then "maintenance", detach" and at >> the end "remove"? > > > Yes, these should be your first steps. > >> >> I think I continue to see 4 LUNs at this point, correct? We do not manage devices, so this must be performed manually by the system administrator. To remove devices, you have to: 1. On the storage server, unzone the devices so the host(s) will not see them This must be done before you remove the multipath device and paths, otherwise vdsm will discover the device again during the periodic scsi/fc rescans. 2. On all hosts ,remove the multipath device and the underlying scsi devices, as explained here: https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Linux/7/html/Storage_Administration_Guide/removing-path-to-storage-device.html Nir > > > Yes. > >> >> Now I proceed with removal of lun at storage array level? >> >> Should I select an SD line and then "Scan disks" to see refresh the SAN >> and see in multipath only 2 of them at the end? >> Or any manual command at host level before removal from array? > > > After removing the storage domains you should be able to remove the luns. I > am not extremely familiar with the multipath and low-level scsi commands but > I would try the scan disks button and if the luns are not gone from your > host you can manually remove them. I think that involves removing the > device from multipath (multipath -d) and deleting it from the scsi > subsystem. > >> Thanks in advance > > > Hope this helped you. > > > ___ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users > ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] best way to remove SAN lun
On Tue, Feb 21, 2017 at 5:12 PM, Adam Litke wrote: > > > On Tue, Feb 21, 2017 at 10:19 AM, Gianluca Cecchi < > gianluca.cec...@gmail.com> wrote: > >> Hello, >> currently I have a cluster of 3 hosts where each one has FC SAN >> connectivity to 4 LUNs: 3 are already configured as storage domains (1TB, >> 2TB, 4TB), one is free, not allocated. >> See here for screenshot: >> https://drive.google.com/file/d/0BwoPbcrMv8mvRVZZMTlNcTQ5MGs >> /view?usp=sharing >> >> At the moment the command "multipath -l" run on hosts shows all the 4 >> LUNs. >> >> Now I want to do 2 things at storage array level: >> >> - remove the 2TB storage domain LUN >> - remove the 20Gb LUN not yet allocated >> >> What is the correct workflow, supposing I have already emptied the 2TB >> from VM disks ad such? >> Select 2Tb SD, then Datacenter subtab, then "maintenance", detach" and at >> the end "remove"? >> > > Yes, these should be your first steps. > > >> I think I continue to see 4 LUNs at this point, correct? >> > > Yes. > > >> Now I proceed with removal of lun at storage array level? >> >> Should I select an SD line and then "Scan disks" to see refresh the SAN >> and see in multipath only 2 of them at the end? >> Or any manual command at host level before removal from array? >> > > After removing the storage domains you should be able to remove the luns. > I am not extremely familiar with the multipath and low-level scsi commands > but I would try the scan disks button and if the luns are not gone from > your host you can manually remove them. I think that involves removing the > device from multipath (multipath -d) and deleting it from the scsi > subsystem. > > Thanks in advance >> > > Hope this helped you. > > Hello, the "Scan Disks" seems related to the particular storage domain selected in storage tab, not overall FC SAN connectivity... If I then select "manage domain", it still shows the now missing disks with an exclamation mark aside I try to follow standard RH EL 7 way for removal: https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Linux/7/html/Storage_Administration_Guide/removing_devices.html I can remove at os level the 20Gb lun that was never used in oVirt, but for the previous 2Tb storage domain related LUN I get the error that is in use [root@ovmsrv06 ~]# multipath -f 3600a0b8000299902cd3c5501458f Feb 21 17:25:58 | 3600a0b8000299902cd3c5501458f: map in use Feb 21 17:25:58 | failed to remove multipath map 3600a0b8000299902cd3c5501458f [root@ovmsrv06 ~]# [root@ovmsrv06 ~]# fuser /dev/mapper/3600a0b8000299902cd3c5501458f [root@ovmsrv06 ~]# [root@ovmsrv06 ~]# ll /dev/mapper/3600a0b8000299902cd3c5501458f lrwxrwxrwx. 1 root root 7 Feb 21 17:25 /dev/mapper/3600a0b8000299902cd3c5501458f -> ../dm-4 [root@ovmsrv06 ~]# fuser /dev/dm-4 [root@ovmsrv06 ~]# Strange thing is that vgs command returns differrent value on th three hosts [root@ovmsrv05 vdsm]# vgs VG #PV #LV #SN Attr VSize VFree 922b5269-ab56-4c4d-838f-49d33427e2ab 1 22 0 wz--n- 4.00t 3.49t cl_ovmsrv051 3 0 wz--n- 67.33g0 [root@ovmsrv05 vdsm]# [root@ovmsrv06 ~]# vgs VG #PV #LV #SN Attr VSize VFree 922b5269-ab56-4c4d-838f-49d33427e2ab 1 22 0 wz--n- 4.00t 3.49t cl 1 3 0 wz--n- 67.33g0 [root@ovmsrv06 ~]# [root@ovmsrv07 vdsm]# vgs VG #PV #LV #SN Attr VSize VFree 900b1853-e192-4661-a0f9-7c7c396f6f49 1 10 0 wz--n- 2.00t 1.76t 922b5269-ab56-4c4d-838f-49d33427e2ab 1 27 0 wz--n- 4.00t 3.34t cl 1 3 0 wz--n- 67.33g0 [root@ovmsrv07 vdsm]# So no host as a VG the 1TB storage domain and In particular ovmsrv07 has a VG of 2TB that I suspect was the previosu storage domain [root@ovmsrv07 vdsm]# lvs 900b1853-e192-4661-a0f9-7c7c396f6f49 LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert 35b8834e-a429-4223-b293-51d562b6def4 900b1853-e192-4661-a0f9-7c7c396f6f49 -wi--- 128.00m 7ed43974-1039-4a68-a8b3-321e7594fe4c 900b1853-e192-4661-a0f9-7c7c396f6f49 -wi--- 240.00g d7f6be37-0f6c-43e3-b0af-a511fc59c842 900b1853-e192-4661-a0f9-7c7c396f6f49 -wi--- 128.00m ids 900b1853-e192-4661-a0f9-7c7c396f6f49 -wi-a- 128.00m inbox900b1853-e192-4661-a0f9-7c7c396f6f49 -wi-a- 128.00m leases 900b1853-e192-4661-a0f9-7c7c396f6f49 -wi-a- 2.00g master 900b1853-e192-4661-a0f9-7c7c396f6f49 -wi-a- 1.00g metadata 900b1853-e192-4661-a0f9-7c7c396f6f49 -wi-a- 512.00m outbox 900b1853-e192-4661-a0f9-7c7c396f6f49 -wi-a- 128.00m xleases 900b1853-e192-4661-a0f9-7c7c396f6f49 -wi
Re: [ovirt-users] best way to remove SAN lun
On Tue, Feb 21, 2017 at 10:19 AM, Gianluca Cecchi wrote: > Hello, > currently I have a cluster of 3 hosts where each one has FC SAN > connectivity to 4 LUNs: 3 are already configured as storage domains (1TB, > 2TB, 4TB), one is free, not allocated. > See here for screenshot: > https://drive.google.com/file/d/0BwoPbcrMv8mvRVZZMTlNcTQ5MGs/ > view?usp=sharing > > At the moment the command "multipath -l" run on hosts shows all the 4 LUNs. > > Now I want to do 2 things at storage array level: > > - remove the 2TB storage domain LUN > - remove the 20Gb LUN not yet allocated > > What is the correct workflow, supposing I have already emptied the 2TB > from VM disks ad such? > Select 2Tb SD, then Datacenter subtab, then "maintenance", detach" and at > the end "remove"? > Yes, these should be your first steps. > I think I continue to see 4 LUNs at this point, correct? > Yes. > Now I proceed with removal of lun at storage array level? > > Should I select an SD line and then "Scan disks" to see refresh the SAN > and see in multipath only 2 of them at the end? > Or any manual command at host level before removal from array? > After removing the storage domains you should be able to remove the luns. I am not extremely familiar with the multipath and low-level scsi commands but I would try the scan disks button and if the luns are not gone from your host you can manually remove them. I think that involves removing the device from multipath (multipath -d) and deleting it from the scsi subsystem. Thanks in advance > Hope this helped you. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users