About 12TB altogether.

> On Mar 21, 2018, at 5:21 PM, r...@italy1.com wrote:
> 
> How much space do you have?
> 
>  dal mio iPhone X 
> 
> Il giorno 21 mar 2018, alle ore 17:10, Father Vlasie <fv@spots.school 
> <mailto:fv@spots.school>> ha scritto:
> 
>> Yes, I agree, it does seem to be an LVM issue rather than cinder. I will 
>> pursue that course.
>> 
>> Thank you all for your help, it is fantastic having a support mailing list 
>> like this!
>> 
>> FV
>> 
>>> On Mar 21, 2018, at 4:45 AM, Vagner Farias <vfar...@redhat.com 
>>> <mailto:vfar...@redhat.com>> wrote:
>>> 
>>> It seems your LVM thin pool metadata is corrupt. I'm not familiar with this 
>>> issue and can't guide you on how to fix it. Although this could have been 
>>> caused by cinder, it's an LVM issue and if you don't get more answers here 
>>> you may try some Linux related forum. 
>>> 
>>> On a quick search on "lvm2 thinpool metadata mismatch" I could find several 
>>> possible causes and solution paths. 
>>> 
>>> I hope that helps. 
>>> 
>>> Vagner Farias
>>> 
>>> 
>>> Em ter, 20 de mar de 2018 22:29, Father Vlasie <fv@spots.school 
>>> <mailto:fv@spots.school>> escreveu:
>>> Your help is much appreciated! Thank you.
>>> 
>>> The cinder service is running on the controller node and it is using a disk 
>>> partition not the loopback device, I did change the default configuration 
>>> during install with PackStack.
>>> 
>>> [root@plato ~]# pvs
>>>   PV         VG             Fmt  Attr PSize    PFree
>>>   /dev/vda3  centos         lvm2 a--  1022.80g    4.00m
>>>   /dev/vdb1  cinder-volumes lvm2 a--   <10.00t <511.85g
>>> 
>>> [root@plato ~]# lvchange -a y volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5
>>>   Volume group "volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5" not found
>>>   Cannot process volume group volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5
>>> 
>>> [root@plato ~]# lvchange -a y cinder-volumes
>>>   Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>> transaction_id is 0, while expected 72.
>>>   Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>> transaction_id is 0, while expected 72.
>>>   Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>> transaction_id is 0, while expected 72.
>>>   Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>> transaction_id is 0, while expected 72.
>>>   Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>> transaction_id is 0, while expected 72.
>>>   Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>> transaction_id is 0, while expected 72.
>>>   Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>> transaction_id is 0, while expected 72.
>>> 
>>> 
>>> 
>>> 
>>> > On Mar 20, 2018, at 6:05 PM, Vagner Farias <vfar...@redhat.com 
>>> > <mailto:vfar...@redhat.com>> wrote:
>>> >
>>> > Will "lvchange -a y lvname" activate it?
>>> >
>>> > If not, considering that you're using Pike on Centos, there's a chance 
>>> > you may be using the cinder-volumes backed by a loopback file.  I guess 
>>> > both packstack & tripleo will configure this by default if you don't 
>>> > change the configuration. At least tripleo won't configure this loopback 
>>> > device to be activated automatically on boot. An option would be to 
>>> > include lines like the following in /etc/rc.d/rc.local:
>>> >
>>> > losetup /dev/loop0 /var/lib/cinder/cinder-volumes
>>> > vgscan
>>> >
>>> > Last but not least, if this is actually the case, I wouldn't recommend 
>>> > using loopback devices for LVM SCSI driver. In fact, if you can use any 
>>> > other driver capable of delivering HA, it'd be better (unless this is 
>>> > some POC or an environment without tight SLAs).
>>> >
>>> > Vagner Farias
>>> >
>>> >
>>> > Em ter, 20 de mar de 2018 21:24, Father Vlasie <fv@spots.school 
>>> > <mailto:fv@spots.school>> escreveu:
>>> > Here is the output of lvdisplay:
>>> >
>>> > [root@plato ~]# lvdisplay
>>> >   --- Logical volume ---
>>> >   LV Name                cinder-volumes-pool
>>> >   VG Name                cinder-volumes
>>> >   LV UUID                PEkGKb-fhAc-CJD2-uDDA-k911-SIX9-1uyvFo
>>> >   LV Write Access        read/write
>>> >   LV Creation host, time plato, 2018-02-01 13:33:51 -0800
>>> >   LV Pool metadata       cinder-volumes-pool_tmeta
>>> >   LV Pool data           cinder-volumes-pool_tdata
>>> >   LV Status              NOT available
>>> >   LV Size                9.50 TiB
>>> >   Current LE             2490368
>>> >   Segments               1
>>> >   Allocation             inherit
>>> >   Read ahead sectors     auto
>>> >
>>> >   --- Logical volume ---
>>> >   LV Path                
>>> > /dev/cinder-volumes/volume-8f4a5fff-749f-47fe-976f-6157f58a4d9e
>>> >   LV Name                volume-8f4a5fff-749f-47fe-976f-6157f58a4d9e
>>> >   VG Name                cinder-volumes
>>> >   LV UUID                C2o7UD-uqFp-3L3r-F0Ys-etjp-QBJr-idBhb0
>>> >   LV Write Access        read/write
>>> >   LV Creation host, time plato, 2018-02-02 10:18:41 -0800
>>> >   LV Pool name           cinder-volumes-pool
>>> >   LV Status              NOT available
>>> >   LV Size                1.00 GiB
>>> >   Current LE             256
>>> >   Segments               1
>>> >   Allocation             inherit
>>> >   Read ahead sectors     auto
>>> >
>>> >   --- Logical volume ---
>>> >   LV Path                
>>> > /dev/cinder-volumes/volume-6ad82e98-c8e2-4837-bffd-079cf76afbe3
>>> >   LV Name                volume-6ad82e98-c8e2-4837-bffd-079cf76afbe3
>>> >   VG Name                cinder-volumes
>>> >   LV UUID                qisf80-j4XV-PpFy-f7yt-ZpJS-99v0-m03Ql4
>>> >   LV Write Access        read/write
>>> >   LV Creation host, time plato, 2018-02-02 10:26:46 -0800
>>> >   LV Pool name           cinder-volumes-pool
>>> >   LV Status              NOT available
>>> >   LV Size                1.00 GiB
>>> >   Current LE             256
>>> >   Segments               1
>>> >   Allocation             inherit
>>> >   Read ahead sectors     auto
>>> >
>>> >   --- Logical volume ---
>>> >   LV Path                
>>> > /dev/cinder-volumes/volume-ee107488-2559-4116-aa7b-0da02fd5f693
>>> >   LV Name                volume-ee107488-2559-4116-aa7b-0da02fd5f693
>>> >   VG Name                cinder-volumes
>>> >   LV UUID                FS9Y2o-HYe2-HK03-yM0Z-P7GO-kAzD-cOYNTb
>>> >   LV Write Access        read/write
>>> >   LV Creation host, time plato.spots.onsite, 2018-02-12 10:28:57 -0800
>>> >   LV Pool name           cinder-volumes-pool
>>> >   LV Status              NOT available
>>> >   LV Size                40.00 GiB
>>> >   Current LE             10240
>>> >   Segments               1
>>> >   Allocation             inherit
>>> >   Read ahead sectors     auto
>>> >
>>> >   --- Logical volume ---
>>> >   LV Path                
>>> > /dev/cinder-volumes/volume-d6f0260d-21b5-43e7-afe5-84e0502fa734
>>> >   LV Name                volume-d6f0260d-21b5-43e7-afe5-84e0502fa734
>>> >   VG Name                cinder-volumes
>>> >   LV UUID                b6pX01-mOEH-3j3K-32NJ-OHsz-UMQe-y10vSM
>>> >   LV Write Access        read/write
>>> >   LV Creation host, time plato.spots.onsite, 2018-02-14 14:24:41 -0800
>>> >   LV Pool name           cinder-volumes-pool
>>> >   LV Status              NOT available
>>> >   LV Size                40.00 GiB
>>> >   Current LE             10240
>>> >   Segments               1
>>> >   Allocation             inherit
>>> >   Read ahead sectors     auto
>>> >
>>> >   --- Logical volume ---
>>> >   LV Path                
>>> > /dev/cinder-volumes/volume-a7bd0bc8-8cbc-4053-bdc2-2eb9bfb0f147
>>> >   LV Name                volume-a7bd0bc8-8cbc-4053-bdc2-2eb9bfb0f147
>>> >   VG Name                cinder-volumes
>>> >   LV UUID                T07JAE-3CNU-CpwN-BUEr-aAJG-VxP5-1qFYZz
>>> >   LV Write Access        read/write
>>> >   LV Creation host, time plato.spots.onsite, 2018-03-12 10:33:24 -0700
>>> >   LV Pool name           cinder-volumes-pool
>>> >   LV Status              NOT available
>>> >   LV Size                4.00 GiB
>>> >   Current LE             1024
>>> >   Segments               1
>>> >   Allocation             inherit
>>> >   Read ahead sectors     auto
>>> >
>>> >   --- Logical volume ---
>>> >   LV Path                
>>> > /dev/cinder-volumes/volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5
>>> >   LV Name                volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5
>>> >   VG Name                cinder-volumes
>>> >   LV UUID                IB0q1n-NnkR-tx5w-BbBu-LamG-jCbQ-mYXWyC
>>> >   LV Write Access        read/write
>>> >   LV Creation host, time plato.spots.onsite, 2018-03-14 09:52:14 -0700
>>> >   LV Pool name           cinder-volumes-pool
>>> >   LV Status              NOT available
>>> >   LV Size                40.00 GiB
>>> >   Current LE             10240
>>> >   Segments               1
>>> >   Allocation             inherit
>>> >   Read ahead sectors     auto
>>> >
>>> >   --- Logical volume ---
>>> >   LV Path                /dev/centos/root
>>> >   LV Name                root
>>> >   VG Name                centos
>>> >   LV UUID                nawE4n-dOHs-VsNH-f9hL-te05-mvGC-WoFQzv
>>> >   LV Write Access        read/write
>>> >   LV Creation host, time localhost, 2018-01-22 09:50:38 -0800
>>> >   LV Status              available
>>> >   # open                 1
>>> >   LV Size                50.00 GiB
>>> >   Current LE             12800
>>> >   Segments               1
>>> >   Allocation             inherit
>>> >   Read ahead sectors     auto
>>> >   - currently set to     8192
>>> >   Block device           253:0
>>> >
>>> >   --- Logical volume ---
>>> >   LV Path                /dev/centos/swap
>>> >   LV Name                swap
>>> >   VG Name                centos
>>> >   LV UUID                Vvlni4-nwTl-ORwW-Gg8b-5y4h-kXJ5-T67cKU
>>> >   LV Write Access        read/write
>>> >   LV Creation host, time localhost, 2018-01-22 09:50:38 -0800
>>> >   LV Status              available
>>> >   # open                 2
>>> >   LV Size                8.12 GiB
>>> >   Current LE             2080
>>> >   Segments               1
>>> >   Allocation             inherit
>>> >   Read ahead sectors     auto
>>> >   - currently set to     8192
>>> >   Block device           253:1
>>> >
>>> >   --- Logical volume ---
>>> >   LV Path                /dev/centos/home
>>> >   LV Name                home
>>> >   VG Name                centos
>>> >   LV UUID                lCXJ7v-jeOC-DFKI-unXa-HUKx-9DXp-nmzSMg
>>> >   LV Write Access        read/write
>>> >   LV Creation host, time localhost, 2018-01-22 09:50:39 -0800
>>> >   LV Status              available
>>> >   # open                 1
>>> >   LV Size                964.67 GiB
>>> >   Current LE             246956
>>> >   Segments               1
>>> >   Allocation             inherit
>>> >   Read ahead sectors     auto
>>> >   - currently set to     8192
>>> >   Block device           253:2
>>> >
>>> >
>>> > > On Mar 20, 2018, at 4:51 PM, Remo Mattei <r...@italy1.com 
>>> > > <mailto:r...@italy1.com>> wrote:
>>> > >
>>> > > I think you need to provide a bit of additional info. Did you look at 
>>> > > the logs? What version of os are you running? Etc.
>>> > >
>>> > > Inviato da iPhone
>>> > >
>>> > >> Il giorno 20 mar 2018, alle ore 16:15, Father Vlasie <fv@spots.school 
>>> > >> <mailto:fv@spots.school>> ha scritto:
>>> > >>
>>> > >> Hello everyone,
>>> > >>
>>> > >> I am in need of help with my Cinder volumes which have all become 
>>> > >> unavailable.
>>> > >>
>>> > >> Is there anyone who would be willing to log in to my system and have a 
>>> > >> look?
>>> > >>
>>> > >> My cinder volumes are listed as "NOT available" and my attempts to 
>>> > >> mount them have been in vain. I have tried: vgchange -a y
>>> > >>
>>> > >> with result showing as:  0 logical volume(s) in volume group 
>>> > >> "cinder-volumes" now active
>>> > >>
>>> > >> I am a bit desperate because some of the data is critical and, I am 
>>> > >> ashamed to say, I do not have a backup.
>>> > >>
>>> > >> Any help or suggestions would be very much appreciated.
>>> > >>
>>> > >> FV
>>> > >> _______________________________________________
>>> > >> Mailing list: 
>>> > >> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack 
>>> > >> <http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack>
>>> > >> Post to     : openstack@lists.openstack.org 
>>> > >> <mailto:openstack@lists.openstack.org>
>>> > >> Unsubscribe : 
>>> > >> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack 
>>> > >> <http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack>
>>> > >
>>> >
>>> >
>>> > _______________________________________________
>>> > Mailing list: 
>>> > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack 
>>> > <http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack>
>>> > Post to     : openstack@lists.openstack.org 
>>> > <mailto:openstack@lists.openstack.org>
>>> > Unsubscribe : 
>>> > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack 
>>> > <http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack>
>>> 
>> 

_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to     : openstack@lists.openstack.org
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack

Reply via email to