error: unrecognized arguments: --debug

[root@plato ~(keystone_admin)]# cinder create --display_name TEST 1
+--------------------------------+--------------------------------------+
| Property                       | Value                                |
+--------------------------------+--------------------------------------+
| attachments                    | []                                   |
| availability_zone              | nova                                 |
| bootable                       | false                                |
| consistencygroup_id            | None                                 |
| created_at                     | 2018-03-22T00:57:49.000000           |
| description                    | None                                 |
| encrypted                      | False                                |
| id                             | 6c42d4dc-edff-4803-a2fa-1efc535927fa |
| metadata                       | {}                                   |
| migration_status               | None                                 |
| multiattach                    | False                                |
| name                           | TEST                                 |
| os-vol-host-attr:host          | None                                 |
| os-vol-mig-status-attr:migstat | None                                 |
| os-vol-mig-status-attr:name_id | None                                 |
| os-vol-tenant-attr:tenant_id   | e1ceb67d89314c01add05a0086772df3     |
| replication_status             | None                                 |
| size                           | 1                                    |
| snapshot_id                    | None                                 |
| source_volid                   | None                                 |
| status                         | creating                             |
| updated_at                     | None                                 |
| user_id                        | ced549e6e1b345be889e11b1c16cf6d9     |
| volume_type                    | iscsi                                |
+--------------------------------+--------------------------------------+

[root@plato ~(keystone_admin)]# less /var/log/cinder/volume.log 

2018-03-21 18:00:21.703 2572 WARNING cinder.volume.manager 
[req-b519dd31-2a7a-4188-835a-1a6d3ea9b7b0 - - - - -] Update driver status 
failed: (config name lvm) is uninitialized.
2018-03-21 18:00:28.268 2572 ERROR cinder.service [-] Manager for service 
cinder-volume plato.spots.onsite@lvm is reporting problems, not sending 
heartbeat. Service will appear “down”.

The same messages repeat for a long time.

lvdisplay does not show the new volume.

Thank you,
FV

> On Mar 21, 2018, at 5:35 PM, r...@italy1.com wrote:
> 
> Ok the pool is ok looks like you have several volumes probably vms. Did you 
> check targetcli? I cannot remember what your cinder-volume says in the log. 
> Can you try to create a volume cinder create 1 and look at the log and see 
> what the error is add —debug to the command its a dash dash somehow my iPhone 
> converted it 
> 
>  dal mio iPhone X 
> 
> Il giorno 21 mar 2018, alle ore 17:28, Father Vlasie <fv@spots.school> ha 
> scritto:
> 
>> [root@plato ~]# lvdisplay
>>   --- Logical volume ---
>>   LV Name                cinder-volumes-pool
>>   VG Name                cinder-volumes
>>   LV UUID                PEkGKb-fhAc-CJD2-uDDA-k911-SIX9-1uyvFo
>>   LV Write Access        read/write
>>   LV Creation host, time plato, 2018-02-01 13:33:51 -0800
>>   LV Pool metadata       cinder-volumes-pool_tmeta
>>   LV Pool data           cinder-volumes-pool_tdata
>>   LV Status              NOT available
>>   LV Size                9.50 TiB
>>   Current LE             2490368
>>   Segments               1
>>   Allocation             inherit
>>   Read ahead sectors     auto
>>    
>>   --- Logical volume ---
>>   LV Path                
>> /dev/cinder-volumes/volume-8f4a5fff-749f-47fe-976f-6157f58a4d9e
>>   LV Name                volume-8f4a5fff-749f-47fe-976f-6157f58a4d9e
>>   VG Name                cinder-volumes
>>   LV UUID                C2o7UD-uqFp-3L3r-F0Ys-etjp-QBJr-idBhb0
>>   LV Write Access        read/write
>>   LV Creation host, time plato, 2018-02-02 10:18:41 -0800
>>   LV Pool name           cinder-volumes-pool
>>   LV Status              NOT available
>>   LV Size                1.00 GiB
>>   Current LE             256
>>   Segments               1
>>   Allocation             inherit
>>   Read ahead sectors     auto
>>    
>>   --- Logical volume ---
>>   LV Path                
>> /dev/cinder-volumes/volume-6ad82e98-c8e2-4837-bffd-079cf76afbe3
>>   LV Name                volume-6ad82e98-c8e2-4837-bffd-079cf76afbe3
>>   VG Name                cinder-volumes
>>   LV UUID                qisf80-j4XV-PpFy-f7yt-ZpJS-99v0-m03Ql4
>>   LV Write Access        read/write
>>   LV Creation host, time plato, 2018-02-02 10:26:46 -0800
>>   LV Pool name           cinder-volumes-pool
>>   LV Status              NOT available
>>   LV Size                1.00 GiB
>>   Current LE             256
>>   Segments               1
>>   Allocation             inherit
>>   Read ahead sectors     auto
>>    
>>   --- Logical volume ---
>>   LV Path                
>> /dev/cinder-volumes/volume-ee107488-2559-4116-aa7b-0da02fd5f693
>>   LV Name                volume-ee107488-2559-4116-aa7b-0da02fd5f693
>>   VG Name                cinder-volumes
>>   LV UUID                FS9Y2o-HYe2-HK03-yM0Z-P7GO-kAzD-cOYNTb
>>   LV Write Access        read/write
>>   LV Creation host, time plato.spots.onsite, 2018-02-12 10:28:57 -0800
>>   LV Pool name           cinder-volumes-pool
>>   LV Status              NOT available
>>   LV Size                40.00 GiB
>>   Current LE             10240
>>   Segments               1
>>   Allocation             inherit
>>   Read ahead sectors     auto
>>    
>>   --- Logical volume ---
>>   LV Path                
>> /dev/cinder-volumes/volume-d6f0260d-21b5-43e7-afe5-84e0502fa734
>>   LV Name                volume-d6f0260d-21b5-43e7-afe5-84e0502fa734
>>   VG Name                cinder-volumes
>>   LV UUID                b6pX01-mOEH-3j3K-32NJ-OHsz-UMQe-y10vSM
>>   LV Write Access        read/write
>>   LV Creation host, time plato.spots.onsite, 2018-02-14 14:24:41 -0800
>>   LV Pool name           cinder-volumes-pool
>>   LV Status              NOT available
>>   LV Size                40.00 GiB
>>   Current LE             10240
>>   Segments               1
>>   Allocation             inherit
>>   Read ahead sectors     auto
>>    
>>   --- Logical volume ---
>>   LV Path                
>> /dev/cinder-volumes/volume-a7bd0bc8-8cbc-4053-bdc2-2eb9bfb0f147
>>   LV Name                volume-a7bd0bc8-8cbc-4053-bdc2-2eb9bfb0f147
>>   VG Name                cinder-volumes
>>   LV UUID                T07JAE-3CNU-CpwN-BUEr-aAJG-VxP5-1qFYZz
>>   LV Write Access        read/write
>>   LV Creation host, time plato.spots.onsite, 2018-03-12 10:33:24 -0700
>>   LV Pool name           cinder-volumes-pool
>>   LV Status              NOT available
>>   LV Size                4.00 GiB
>>   Current LE             1024
>>   Segments               1
>>   Allocation             inherit
>>   Read ahead sectors     auto
>>    
>>   --- Logical volume ---
>>   LV Path                
>> /dev/cinder-volumes/volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5
>>   LV Name                volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5
>>   VG Name                cinder-volumes
>>   LV UUID                IB0q1n-NnkR-tx5w-BbBu-LamG-jCbQ-mYXWyC
>>   LV Write Access        read/write
>>   LV Creation host, time plato.spots.onsite, 2018-03-14 09:52:14 -0700
>>   LV Pool name           cinder-volumes-pool
>>   LV Status              NOT available
>>   LV Size                40.00 GiB
>>   Current LE             10240
>>   Segments               1
>>   Allocation             inherit
>>   Read ahead sectors     auto
>>    
>>   --- Logical volume ---
>>   LV Path                /dev/centos/root
>>   LV Name                root
>>   VG Name                centos
>>   LV UUID                nawE4n-dOHs-VsNH-f9hL-te05-mvGC-WoFQzv
>>   LV Write Access        read/write
>>   LV Creation host, time localhost, 2018-01-22 09:50:38 -0800
>>   LV Status              available
>>   # open                 1
>>   LV Size                50.00 GiB
>>   Current LE             12800
>>   Segments               1
>>   Allocation             inherit
>>   Read ahead sectors     auto
>>   - currently set to     8192
>>   Block device           253:0
>>    
>>   --- Logical volume ---
>>   LV Path                /dev/centos/swap
>>   LV Name                swap
>>   VG Name                centos
>>   LV UUID                Vvlni4-nwTl-ORwW-Gg8b-5y4h-kXJ5-T67cKU
>>   LV Write Access        read/write
>>   LV Creation host, time localhost, 2018-01-22 09:50:38 -0800
>>   LV Status              available
>>   # open                 2
>>   LV Size                8.12 GiB
>>   Current LE             2080
>>   Segments               1
>>   Allocation             inherit
>>   Read ahead sectors     auto
>>   - currently set to     8192
>>   Block device           253:1
>>    
>>   --- Logical volume ---
>>   LV Path                /dev/centos/home
>>   LV Name                home
>>   VG Name                centos
>>   LV UUID                lCXJ7v-jeOC-DFKI-unXa-HUKx-9DXp-nmzSMg
>>   LV Write Access        read/write
>>   LV Creation host, time localhost, 2018-01-22 09:50:39 -0800
>>   LV Status              available
>>   # open                 1
>>   LV Size                964.67 GiB
>>   Current LE             246956
>>   Segments               1
>>   Allocation             inherit
>>   Read ahead sectors     auto
>>   - currently set to     8192
>>   Block device           253:2
>>    
>> 
>> 
>>> On Mar 21, 2018, at 5:25 PM, r...@italy1.com wrote:
>>> 
>>> Can you do an lvdisplay 
>>> 
>>>  dal mio iPhone X 
>>> 
>>> Il giorno 21 mar 2018, alle ore 17:23, Father Vlasie <fv@spots.school> ha 
>>> scritto:
>>> 
>>>> About 12TB altogether.
>>>> 
>>>>> On Mar 21, 2018, at 5:21 PM, r...@italy1.com wrote:
>>>>> 
>>>>> How much space do you have?
>>>>> 
>>>>>  dal mio iPhone X 
>>>>> 
>>>>> Il giorno 21 mar 2018, alle ore 17:10, Father Vlasie <fv@spots.school> ha 
>>>>> scritto:
>>>>> 
>>>>>> Yes, I agree, it does seem to be an LVM issue rather than cinder. I will 
>>>>>> pursue that course.
>>>>>> 
>>>>>> Thank you all for your help, it is fantastic having a support mailing 
>>>>>> list like this!
>>>>>> 
>>>>>> FV
>>>>>> 
>>>>>>> On Mar 21, 2018, at 4:45 AM, Vagner Farias <vfar...@redhat.com> wrote:
>>>>>>> 
>>>>>>> It seems your LVM thin pool metadata is corrupt. I'm not familiar with 
>>>>>>> this issue and can't guide you on how to fix it. Although this could 
>>>>>>> have been caused by cinder, it's an LVM issue and if you don't get more 
>>>>>>> answers here you may try some Linux related forum. 
>>>>>>> 
>>>>>>> On a quick search on "lvm2 thinpool metadata mismatch" I could find 
>>>>>>> several possible causes and solution paths. 
>>>>>>> 
>>>>>>> I hope that helps. 
>>>>>>> 
>>>>>>> Vagner Farias
>>>>>>> 
>>>>>>> 
>>>>>>> Em ter, 20 de mar de 2018 22:29, Father Vlasie <fv@spots.school> 
>>>>>>> escreveu:
>>>>>>> Your help is much appreciated! Thank you.
>>>>>>> 
>>>>>>> The cinder service is running on the controller node and it is using a 
>>>>>>> disk partition not the loopback device, I did change the default 
>>>>>>> configuration during install with PackStack.
>>>>>>> 
>>>>>>> [root@plato ~]# pvs
>>>>>>>   PV         VG             Fmt  Attr PSize    PFree
>>>>>>>   /dev/vda3  centos         lvm2 a--  1022.80g    4.00m
>>>>>>>   /dev/vdb1  cinder-volumes lvm2 a--   <10.00t <511.85g
>>>>>>> 
>>>>>>> [root@plato ~]# lvchange -a y 
>>>>>>> volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5
>>>>>>>   Volume group "volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5" not found
>>>>>>>   Cannot process volume group 
>>>>>>> volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5
>>>>>>> 
>>>>>>> [root@plato ~]# lvchange -a y cinder-volumes
>>>>>>>   Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>>>>>> transaction_id is 0, while expected 72.
>>>>>>>   Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>>>>>> transaction_id is 0, while expected 72.
>>>>>>>   Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>>>>>> transaction_id is 0, while expected 72.
>>>>>>>   Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>>>>>> transaction_id is 0, while expected 72.
>>>>>>>   Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>>>>>> transaction_id is 0, while expected 72.
>>>>>>>   Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>>>>>> transaction_id is 0, while expected 72.
>>>>>>>   Thin pool cinder--volumes-cinder--volumes--pool-tpool (253:5) 
>>>>>>> transaction_id is 0, while expected 72.
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> > On Mar 20, 2018, at 6:05 PM, Vagner Farias <vfar...@redhat.com> wrote:
>>>>>>> >
>>>>>>> > Will "lvchange -a y lvname" activate it?
>>>>>>> >
>>>>>>> > If not, considering that you're using Pike on Centos, there's a 
>>>>>>> > chance you may be using the cinder-volumes backed by a loopback file. 
>>>>>>> >  I guess both packstack & tripleo will configure this by default if 
>>>>>>> > you don't change the configuration. At least tripleo won't configure 
>>>>>>> > this loopback device to be activated automatically on boot. An option 
>>>>>>> > would be to include lines like the following in /etc/rc.d/rc.local:
>>>>>>> >
>>>>>>> > losetup /dev/loop0 /var/lib/cinder/cinder-volumes
>>>>>>> > vgscan
>>>>>>> >
>>>>>>> > Last but not least, if this is actually the case, I wouldn't 
>>>>>>> > recommend using loopback devices for LVM SCSI driver. In fact, if you 
>>>>>>> > can use any other driver capable of delivering HA, it'd be better 
>>>>>>> > (unless this is some POC or an environment without tight SLAs).
>>>>>>> >
>>>>>>> > Vagner Farias
>>>>>>> >
>>>>>>> >
>>>>>>> > Em ter, 20 de mar de 2018 21:24, Father Vlasie <fv@spots.school> 
>>>>>>> > escreveu:
>>>>>>> > Here is the output of lvdisplay:
>>>>>>> >
>>>>>>> > [root@plato ~]# lvdisplay
>>>>>>> >   --- Logical volume ---
>>>>>>> >   LV Name                cinder-volumes-pool
>>>>>>> >   VG Name                cinder-volumes
>>>>>>> >   LV UUID                PEkGKb-fhAc-CJD2-uDDA-k911-SIX9-1uyvFo
>>>>>>> >   LV Write Access        read/write
>>>>>>> >   LV Creation host, time plato, 2018-02-01 13:33:51 -0800
>>>>>>> >   LV Pool metadata       cinder-volumes-pool_tmeta
>>>>>>> >   LV Pool data           cinder-volumes-pool_tdata
>>>>>>> >   LV Status              NOT available
>>>>>>> >   LV Size                9.50 TiB
>>>>>>> >   Current LE             2490368
>>>>>>> >   Segments               1
>>>>>>> >   Allocation             inherit
>>>>>>> >   Read ahead sectors     auto
>>>>>>> >
>>>>>>> >   --- Logical volume ---
>>>>>>> >   LV Path                
>>>>>>> > /dev/cinder-volumes/volume-8f4a5fff-749f-47fe-976f-6157f58a4d9e
>>>>>>> >   LV Name                volume-8f4a5fff-749f-47fe-976f-6157f58a4d9e
>>>>>>> >   VG Name                cinder-volumes
>>>>>>> >   LV UUID                C2o7UD-uqFp-3L3r-F0Ys-etjp-QBJr-idBhb0
>>>>>>> >   LV Write Access        read/write
>>>>>>> >   LV Creation host, time plato, 2018-02-02 10:18:41 -0800
>>>>>>> >   LV Pool name           cinder-volumes-pool
>>>>>>> >   LV Status              NOT available
>>>>>>> >   LV Size                1.00 GiB
>>>>>>> >   Current LE             256
>>>>>>> >   Segments               1
>>>>>>> >   Allocation             inherit
>>>>>>> >   Read ahead sectors     auto
>>>>>>> >
>>>>>>> >   --- Logical volume ---
>>>>>>> >   LV Path                
>>>>>>> > /dev/cinder-volumes/volume-6ad82e98-c8e2-4837-bffd-079cf76afbe3
>>>>>>> >   LV Name                volume-6ad82e98-c8e2-4837-bffd-079cf76afbe3
>>>>>>> >   VG Name                cinder-volumes
>>>>>>> >   LV UUID                qisf80-j4XV-PpFy-f7yt-ZpJS-99v0-m03Ql4
>>>>>>> >   LV Write Access        read/write
>>>>>>> >   LV Creation host, time plato, 2018-02-02 10:26:46 -0800
>>>>>>> >   LV Pool name           cinder-volumes-pool
>>>>>>> >   LV Status              NOT available
>>>>>>> >   LV Size                1.00 GiB
>>>>>>> >   Current LE             256
>>>>>>> >   Segments               1
>>>>>>> >   Allocation             inherit
>>>>>>> >   Read ahead sectors     auto
>>>>>>> >
>>>>>>> >   --- Logical volume ---
>>>>>>> >   LV Path                
>>>>>>> > /dev/cinder-volumes/volume-ee107488-2559-4116-aa7b-0da02fd5f693
>>>>>>> >   LV Name                volume-ee107488-2559-4116-aa7b-0da02fd5f693
>>>>>>> >   VG Name                cinder-volumes
>>>>>>> >   LV UUID                FS9Y2o-HYe2-HK03-yM0Z-P7GO-kAzD-cOYNTb
>>>>>>> >   LV Write Access        read/write
>>>>>>> >   LV Creation host, time plato.spots.onsite, 2018-02-12 10:28:57 -0800
>>>>>>> >   LV Pool name           cinder-volumes-pool
>>>>>>> >   LV Status              NOT available
>>>>>>> >   LV Size                40.00 GiB
>>>>>>> >   Current LE             10240
>>>>>>> >   Segments               1
>>>>>>> >   Allocation             inherit
>>>>>>> >   Read ahead sectors     auto
>>>>>>> >
>>>>>>> >   --- Logical volume ---
>>>>>>> >   LV Path                
>>>>>>> > /dev/cinder-volumes/volume-d6f0260d-21b5-43e7-afe5-84e0502fa734
>>>>>>> >   LV Name                volume-d6f0260d-21b5-43e7-afe5-84e0502fa734
>>>>>>> >   VG Name                cinder-volumes
>>>>>>> >   LV UUID                b6pX01-mOEH-3j3K-32NJ-OHsz-UMQe-y10vSM
>>>>>>> >   LV Write Access        read/write
>>>>>>> >   LV Creation host, time plato.spots.onsite, 2018-02-14 14:24:41 -0800
>>>>>>> >   LV Pool name           cinder-volumes-pool
>>>>>>> >   LV Status              NOT available
>>>>>>> >   LV Size                40.00 GiB
>>>>>>> >   Current LE             10240
>>>>>>> >   Segments               1
>>>>>>> >   Allocation             inherit
>>>>>>> >   Read ahead sectors     auto
>>>>>>> >
>>>>>>> >   --- Logical volume ---
>>>>>>> >   LV Path                
>>>>>>> > /dev/cinder-volumes/volume-a7bd0bc8-8cbc-4053-bdc2-2eb9bfb0f147
>>>>>>> >   LV Name                volume-a7bd0bc8-8cbc-4053-bdc2-2eb9bfb0f147
>>>>>>> >   VG Name                cinder-volumes
>>>>>>> >   LV UUID                T07JAE-3CNU-CpwN-BUEr-aAJG-VxP5-1qFYZz
>>>>>>> >   LV Write Access        read/write
>>>>>>> >   LV Creation host, time plato.spots.onsite, 2018-03-12 10:33:24 -0700
>>>>>>> >   LV Pool name           cinder-volumes-pool
>>>>>>> >   LV Status              NOT available
>>>>>>> >   LV Size                4.00 GiB
>>>>>>> >   Current LE             1024
>>>>>>> >   Segments               1
>>>>>>> >   Allocation             inherit
>>>>>>> >   Read ahead sectors     auto
>>>>>>> >
>>>>>>> >   --- Logical volume ---
>>>>>>> >   LV Path                
>>>>>>> > /dev/cinder-volumes/volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5
>>>>>>> >   LV Name                volume-29fa3b6d-1cbf-40db-82bb-1756c6fac9a5
>>>>>>> >   VG Name                cinder-volumes
>>>>>>> >   LV UUID                IB0q1n-NnkR-tx5w-BbBu-LamG-jCbQ-mYXWyC
>>>>>>> >   LV Write Access        read/write
>>>>>>> >   LV Creation host, time plato.spots.onsite, 2018-03-14 09:52:14 -0700
>>>>>>> >   LV Pool name           cinder-volumes-pool
>>>>>>> >   LV Status              NOT available
>>>>>>> >   LV Size                40.00 GiB
>>>>>>> >   Current LE             10240
>>>>>>> >   Segments               1
>>>>>>> >   Allocation             inherit
>>>>>>> >   Read ahead sectors     auto
>>>>>>> >
>>>>>>> >   --- Logical volume ---
>>>>>>> >   LV Path                /dev/centos/root
>>>>>>> >   LV Name                root
>>>>>>> >   VG Name                centos
>>>>>>> >   LV UUID                nawE4n-dOHs-VsNH-f9hL-te05-mvGC-WoFQzv
>>>>>>> >   LV Write Access        read/write
>>>>>>> >   LV Creation host, time localhost, 2018-01-22 09:50:38 -0800
>>>>>>> >   LV Status              available
>>>>>>> >   # open                 1
>>>>>>> >   LV Size                50.00 GiB
>>>>>>> >   Current LE             12800
>>>>>>> >   Segments               1
>>>>>>> >   Allocation             inherit
>>>>>>> >   Read ahead sectors     auto
>>>>>>> >   - currently set to     8192
>>>>>>> >   Block device           253:0
>>>>>>> >
>>>>>>> >   --- Logical volume ---
>>>>>>> >   LV Path                /dev/centos/swap
>>>>>>> >   LV Name                swap
>>>>>>> >   VG Name                centos
>>>>>>> >   LV UUID                Vvlni4-nwTl-ORwW-Gg8b-5y4h-kXJ5-T67cKU
>>>>>>> >   LV Write Access        read/write
>>>>>>> >   LV Creation host, time localhost, 2018-01-22 09:50:38 -0800
>>>>>>> >   LV Status              available
>>>>>>> >   # open                 2
>>>>>>> >   LV Size                8.12 GiB
>>>>>>> >   Current LE             2080
>>>>>>> >   Segments               1
>>>>>>> >   Allocation             inherit
>>>>>>> >   Read ahead sectors     auto
>>>>>>> >   - currently set to     8192
>>>>>>> >   Block device           253:1
>>>>>>> >
>>>>>>> >   --- Logical volume ---
>>>>>>> >   LV Path                /dev/centos/home
>>>>>>> >   LV Name                home
>>>>>>> >   VG Name                centos
>>>>>>> >   LV UUID                lCXJ7v-jeOC-DFKI-unXa-HUKx-9DXp-nmzSMg
>>>>>>> >   LV Write Access        read/write
>>>>>>> >   LV Creation host, time localhost, 2018-01-22 09:50:39 -0800
>>>>>>> >   LV Status              available
>>>>>>> >   # open                 1
>>>>>>> >   LV Size                964.67 GiB
>>>>>>> >   Current LE             246956
>>>>>>> >   Segments               1
>>>>>>> >   Allocation             inherit
>>>>>>> >   Read ahead sectors     auto
>>>>>>> >   - currently set to     8192
>>>>>>> >   Block device           253:2
>>>>>>> >
>>>>>>> >
>>>>>>> > > On Mar 20, 2018, at 4:51 PM, Remo Mattei <r...@italy1.com> wrote:
>>>>>>> > >
>>>>>>> > > I think you need to provide a bit of additional info. Did you look 
>>>>>>> > > at the logs? What version of os are you running? Etc.
>>>>>>> > >
>>>>>>> > > Inviato da iPhone
>>>>>>> > >
>>>>>>> > >> Il giorno 20 mar 2018, alle ore 16:15, Father Vlasie 
>>>>>>> > >> <fv@spots.school> ha scritto:
>>>>>>> > >>
>>>>>>> > >> Hello everyone,
>>>>>>> > >>
>>>>>>> > >> I am in need of help with my Cinder volumes which have all become 
>>>>>>> > >> unavailable.
>>>>>>> > >>
>>>>>>> > >> Is there anyone who would be willing to log in to my system and 
>>>>>>> > >> have a look?
>>>>>>> > >>
>>>>>>> > >> My cinder volumes are listed as "NOT available" and my attempts to 
>>>>>>> > >> mount them have been in vain. I have tried: vgchange -a y
>>>>>>> > >>
>>>>>>> > >> with result showing as:  0 logical volume(s) in volume group 
>>>>>>> > >> "cinder-volumes" now active
>>>>>>> > >>
>>>>>>> > >> I am a bit desperate because some of the data is critical and, I 
>>>>>>> > >> am ashamed to say, I do not have a backup.
>>>>>>> > >>
>>>>>>> > >> Any help or suggestions would be very much appreciated.
>>>>>>> > >>
>>>>>>> > >> FV
>>>>>>> > >> _______________________________________________
>>>>>>> > >> Mailing list: 
>>>>>>> > >> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>>> > >> Post to     : openstack@lists.openstack.org
>>>>>>> > >> Unsubscribe : 
>>>>>>> > >> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>>> > >
>>>>>>> >
>>>>>>> >
>>>>>>> > _______________________________________________
>>>>>>> > Mailing list: 
>>>>>>> > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>>> > Post to     : openstack@lists.openstack.org
>>>>>>> > Unsubscribe : 
>>>>>>> > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>>> 
>>>>>> 
>>>> 
>> 


_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to     : openstack@lists.openstack.org
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack

Reply via email to