Looks like your pacemaker is not running check that out! sudo pcs status
> On Mar 20, 2018, at 6:24 PM, Father Vlasie <fv@spots.school> wrote: > > Your help is much appreciated! Thank you. > > The cinder service is running on the controller node and it is using a disk > partition not the loopback device, I did change the default configuration > during install with PackStack. > > [root@plato ~]# pvs > PV VG Fmt Attr PSize PFree > /dev/vda3 centos lvm2 a-- 1022.80g 4.00m > /dev/vdb1 cinder-volumes lvm2 a-- <10.00t <511.85g > > [root@plato ~]# lvchange -a y volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5 > Volume group "volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5" not found > Cannot process volume group volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5 > > [root@plato ~]# lvchange -a y cinder-volumes > Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) transaction_id > is 0, while expected 72. > Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) transaction_id > is 0, while expected 72. > Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) transaction_id > is 0, while expected 72. > Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) transaction_id > is 0, while expected 72. > Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) transaction_id > is 0, while expected 72. > Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) transaction_id > is 0, while expected 72. > Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) transaction_id > is 0, while expected 72. > > > > >> On Mar 20, 2018, at 6:05 PM, Vagner Farias <vfar...@redhat.com> wrote: >> >> Will "lvchange -a y lvname" activate it? >> >> If not, considering that you're using Pike on Centos, there's a chance you >> may be using the cinder-volumes backed by a loopback file. I guess both >> packstack & tripleo will configure this by default if you don't change the >> configuration. At least tripleo won't configure this loopback device to be >> activated automatically on boot. An option would be to include lines like >> the following in /etc/rc.d/rc.local: >> >> losetup /dev/loop0 /var/lib/cinder/cinder-volumes >> vgscan >> >> Last but not least, if this is actually the case, I wouldn't recommend using >> loopback devices for LVM SCSI driver. In fact, if you can use any other >> driver capable of delivering HA, it'd be better (unless this is some POC or >> an environment without tight SLAs). >> >> Vagner Farias >> >> >> Em ter, 20 de mar de 2018 21:24, Father Vlasie <fv@spots.school> escreveu: >> Here is the output of lvdisplay: >> >> [root@plato ~]# lvdisplay >> --- Logical volume --- >> LV Name cinder-volumes-pool >> VG Name cinder-volumes >> LV UUID PEkGKb-fhAc-CJD2-uDDA-k911-SIX9-1uyvFo >> LV Write Access read/write >> LV Creation host, time plato, 2018-02-01 13:33:51 -0800 >> LV Pool metadata cinder-volumes-pool_tmeta >> LV Pool data cinder-volumes-pool_tdata >> LV Status NOT available >> LV Size 9.50 TiB >> Current LE 2490368 >> Segments 1 >> Allocation inherit >> Read ahead sectors auto >> >> --- Logical volume --- >> LV Path >> /dev/cinder-volumes/volume-8f4a5fff-749f-47fe-976f-6157f58a4d9e >> LV Name volume-8f4a5fff-749f-47fe-976f-6157f58a4d9e >> VG Name cinder-volumes >> LV UUID C2o7UD-uqFp-3L3r-F0Ys-etjp-QBJr-idBhb0 >> LV Write Access read/write >> LV Creation host, time plato, 2018-02-02 10:18:41 -0800 >> LV Pool name cinder-volumes-pool >> LV Status NOT available >> LV Size 1.00 GiB >> Current LE 256 >> Segments 1 >> Allocation inherit >> Read ahead sectors auto >> >> --- Logical volume --- >> LV Path >> /dev/cinder-volumes/volume-6ad82e98-c8e2-4837-bffd-079cf76afbe3 >> LV Name volume-6ad82e98-c8e2-4837-bffd-079cf76afbe3 >> VG Name cinder-volumes >> LV UUID qisf80-j4XV-PpFy-f7yt-ZpJS-99v0-m03Ql4 >> LV Write Access read/write >> LV Creation host, time plato, 2018-02-02 10:26:46 -0800 >> LV Pool name cinder-volumes-pool >> LV Status NOT available >> LV Size 1.00 GiB >> Current LE 256 >> Segments 1 >> Allocation inherit >> Read ahead sectors auto >> >> --- Logical volume --- >> LV Path >> /dev/cinder-volumes/volume-ee107488-2559-4116-aa7b-0da02fd5f693 >> LV Name volume-ee107488-2559-4116-aa7b-0da02fd5f693 >> VG Name cinder-volumes >> LV UUID FS9Y2o-HYe2-HK03-yM0Z-P7GO-kAzD-cOYNTb >> LV Write Access read/write >> LV Creation host, time plato.spots.onsite, 2018-02-12 10:28:57 -0800 >> LV Pool name cinder-volumes-pool >> LV Status NOT available >> LV Size 40.00 GiB >> Current LE 10240 >> Segments 1 >> Allocation inherit >> Read ahead sectors auto >> >> --- Logical volume --- >> LV Path >> /dev/cinder-volumes/volume-d6f0260d-21b5-43e7-afe5-84e0502fa734 >> LV Name volume-d6f0260d-21b5-43e7-afe5-84e0502fa734 >> VG Name cinder-volumes >> LV UUID b6pX01-mOEH-3j3K-32NJ-OHsz-UMQe-y10vSM >> LV Write Access read/write >> LV Creation host, time plato.spots.onsite, 2018-02-14 14:24:41 -0800 >> LV Pool name cinder-volumes-pool >> LV Status NOT available >> LV Size 40.00 GiB >> Current LE 10240 >> Segments 1 >> Allocation inherit >> Read ahead sectors auto >> >> --- Logical volume --- >> LV Path >> /dev/cinder-volumes/volume-a7bd0bc8-8cbc-4053-bdc2-2eb9bfb0f147 >> LV Name volume-a7bd0bc8-8cbc-4053-bdc2-2eb9bfb0f147 >> VG Name cinder-volumes >> LV UUID T07JAE-3CNU-CpwN-BUEr-aAJG-VxP5-1qFYZz >> LV Write Access read/write >> LV Creation host, time plato.spots.onsite, 2018-03-12 10:33:24 -0700 >> LV Pool name cinder-volumes-pool >> LV Status NOT available >> LV Size 4.00 GiB >> Current LE 1024 >> Segments 1 >> Allocation inherit >> Read ahead sectors auto >> >> --- Logical volume --- >> LV Path >> /dev/cinder-volumes/volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5 >> LV Name volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5 >> VG Name cinder-volumes >> LV UUID IB0q1n-NnkR-tx5w-BbBu-LamG-jCbQ-mYXWyC >> LV Write Access read/write >> LV Creation host, time plato.spots.onsite, 2018-03-14 09:52:14 -0700 >> LV Pool name cinder-volumes-pool >> LV Status NOT available >> LV Size 40.00 GiB >> Current LE 10240 >> Segments 1 >> Allocation inherit >> Read ahead sectors auto >> >> --- Logical volume --- >> LV Path /dev/centos/root >> LV Name root >> VG Name centos >> LV UUID nawE4n-dOHs-VsNH-f9hL-te05-mvGC-WoFQzv >> LV Write Access read/write >> LV Creation host, time localhost, 2018-01-22 09:50:38 -0800 >> LV Status available >> # open 1 >> LV Size 50.00 GiB >> Current LE 12800 >> Segments 1 >> Allocation inherit >> Read ahead sectors auto >> - currently set to 8192 >> Block device 253:0 >> >> --- Logical volume --- >> LV Path /dev/centos/swap >> LV Name swap >> VG Name centos >> LV UUID Vvlni4-nwTl-ORwW-Gg8b-5y4h-kXJ5-T67cKU >> LV Write Access read/write >> LV Creation host, time localhost, 2018-01-22 09:50:38 -0800 >> LV Status available >> # open 2 >> LV Size 8.12 GiB >> Current LE 2080 >> Segments 1 >> Allocation inherit >> Read ahead sectors auto >> - currently set to 8192 >> Block device 253:1 >> >> --- Logical volume --- >> LV Path /dev/centos/home >> LV Name home >> VG Name centos >> LV UUID lCXJ7v-jeOC-DFKI-unXa-HUKx-9DXp-nmzSMg >> LV Write Access read/write >> LV Creation host, time localhost, 2018-01-22 09:50:39 -0800 >> LV Status available >> # open 1 >> LV Size 964.67 GiB >> Current LE 246956 >> Segments 1 >> Allocation inherit >> Read ahead sectors auto >> - currently set to 8192 >> Block device 253:2 >> >> >>> On Mar 20, 2018, at 4:51 PM, Remo Mattei <r...@italy1.com> wrote: >>> >>> I think you need to provide a bit of additional info. Did you look at the >>> logs? What version of os are you running? Etc. >>> >>> Inviato da iPhone >>> >>>> Il giorno 20 mar 2018, alle ore 16:15, Father Vlasie <fv@spots.school> ha >>>> scritto: >>>> >>>> Hello everyone, >>>> >>>> I am in need of help with my Cinder volumes which have all become >>>> unavailable. >>>> >>>> Is there anyone who would be willing to log in to my system and have a >>>> look? >>>> >>>> My cinder volumes are listed as "NOT available" and my attempts to mount >>>> them have been in vain. I have tried: vgchange -a y >>>> >>>> with result showing as: 0 logical volume(s) in volume group >>>> "cinder-volumes" now active >>>> >>>> I am a bit desperate because some of the data is critical and, I am >>>> ashamed to say, I do not have a backup. >>>> >>>> Any help or suggestions would be very much appreciated. >>>> >>>> FV >>>> _______________________________________________ >>>> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>>> Post to : openstack@lists.openstack.org >>>> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>> >> >> >> _______________________________________________ >> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >> Post to : openstack@lists.openstack.org >> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >
_______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack