Please check your tenant quota and calculate how much free space in Grizzly
cinder quota-show ${tenant-id}
cinder listIf you are using Havana, there is a convenient command “cinder quota-usage” :) 发件人: Guilherme Russi [mailto:[email protected]] 发送时间: 2013年11月18日 21:12 收件人: Razique Mahroua 抄送: openstack 主题: Re: [Openstack] Multiples storages Hello again guys, I'm trying to overcome another issue with cinder, I'm here trying to create e 500 GB volume, this is my disk: pvscan PV /dev/sdb1 VG cinder-volumes-2 lvm2 [931,51 GiB / 531,51 GiB free] Total: 1 [931,51 GiB] / in use: 1 [931,51 GiB] / in no VG: 0 [0 ] But when I try: cinder create --volume_type lvm_one --display_name v2st3-500 500 I get: ERROR: VolumeSizeExceedsAvailableQuota: Requested volume or snapshot exceeds allowed Gigabytes quota Does anybody know where I begin to fix it? The outputs on cinder-api.log are: ERROR [cinder.api.middleware.fault] Caught error: Requested volume or snapshot exceeds allowed Gigabytes quota Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/cinder/api/middleware/fault.py", line 73, in __call__ return req.get_response(self.application) File "/usr/lib/python2.7/dist-packages/webob/request.py", line 1296, in send application, catch_exc_info=False) File "/usr/lib/python2.7/dist-packages/webob/request.py", line 1260, in call_application app_iter = application(self.environ, start_response) File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__ return resp(environ, start_response) File "/usr/lib/python2.7/dist-packages/keystoneclient/middleware/auth_token.py", line 450, in __call__ return self.app(env, start_response) File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__ return resp(environ, start_response) File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__ return resp(environ, start_response) File "/usr/lib/python2.7/dist-packages/routes/middleware.py", line 131, in __call__ response = self.app(environ, start_response) File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__ return resp(environ, start_response) File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 130, in __call__ resp = self.call_func(req, *args, **self.kwargs) File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 195, in call_func return self.func(req, *args, **kwargs) File "/usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py", line 803, in __call__ content_type, body, accept) File "/usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py", line 851, in _process_stack action_result = self.dispatch(meth, request, action_args) File "/usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py", line 927, in dispatch return method(req=request, **action_args) File "/usr/lib/python2.7/dist-packages/cinder/api/v1/volumes.py", line 358, in create **kwargs) File "/usr/lib/python2.7/dist-packages/cinder/volume/api.py", line 165, in create raise exception.VolumeSizeExceedsAvailableQuota() VolumeSizeExceedsAvailableQuota: Requested volume or snapshot exceeds allowed Gigabytes quota Thank you all. 2013/11/15 Razique Mahroua <[email protected]<mailto:[email protected]>> Awesome :) Razique On 14 Nov 2013, at 15:27, Guilherme Russi wrote: That's right, I've stopped the open-iscsi and tgt process, and the lvremove worked. Thank you all. Regards. 2013/11/13 Razique Mahroua <[email protected]<mailto:[email protected]>> Hey :) that means the volume is still in use. (lvopen : 1) make sure it's not by checking the process, qemu-nbd, etc... On 13 Nov 2013, at 4:50, Guilherme Russi wrote: Hello Razique, I'm here opening this thread again, I've done some cinder delete but when I try to create another storeges it returns there's no space to create a new volume. Here is part of my lvdisplay output: Alloc PE / Size 52224 / 204,00 GiB Free PE / Size 19350 / 75,59 GiB And here is my lvdisplay: --- Logical volume --- LV Name /dev/cinder-volumes/volume-06ccd141-91c4-45e4-b21f-595f4a36779b VG Name cinder-volumes LV UUID wdqxVd-GgUQ-21O4-OWlR-sRT3-HvUA-Q8j9kL LV Write Access read/write LV snapshot status source of /dev/cinder-volumes/_snapshot-04e8414e-2c0e-4fc2-8bff-43dd80ecca09 [active] LV Status available open 0 LV Size 10,00 GiB Current LE 2560 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 256 Block device 252:1 --- Logical volume --- LV Name /dev/cinder-volumes/_snapshot-04e8414e-2c0e-4fc2-8bff-43dd80ecca09 VG Name cinder-volumes LV UUID EZz1lC-a8H2-1PlN-pJTN-XAIm-wW0q-qtUQOc LV Write Access read/write LV snapshot status active destination for /dev/cinder-volumes/volume-06ccd141-91c4-45e4-b21f-595f4a36779b LV Status available open 0 LV Size 10,00 GiB Current LE 2560 COW-table size 10,00 GiB COW-table LE 2560 Allocated to snapshot 0,00% Snapshot chunk size 4,00 KiB Segments 1 Allocation inherit Read ahead sectors auto - currently set to 256 Block device 252:3 --- Logical volume --- LV Name /dev/cinder-volumes/volume-ca36920e-938e-4ad1-b9c4-74c1e28abd31 VG Name cinder-volumes LV UUID b40kQV-P8N4-R6jt-k97Z-I2a1-9TXm-5GXqfz LV Write Access read/write LV Status available open 1 LV Size 60,00 GiB Current LE 15360 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 256 Block device 252:4 --- Logical volume --- LV Name /dev/cinder-volumes/volume-70be4f36-10bd-4877-b841-80333ccfe985 VG Name cinder-volumes LV UUID 2YDrMs-BrYo-aQcZ-8AlX-A4La-HET1-9UQ0gV LV Write Access read/write LV Status available open 1 LV Size 1,00 GiB Current LE 256 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 256 Block device 252:5 --- Logical volume --- LV Name /dev/cinder-volumes/volume-00c532bd-91fb-4a38-b340-4389fb7f0ed5 VG Name cinder-volumes LV UUID MfVOuB-5x5A-jne3-H4Ul-4NP8-eI7b-UYSYE7 LV Write Access read/write LV Status available open 0 LV Size 1,00 GiB Current LE 256 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 256 Block device 252:6 --- Logical volume --- LV Name /dev/cinder-volumes/volume-ae133dbc-6141-48cf-beeb-9d6576e57a45 VG Name cinder-volumes LV UUID 53w8j3-WT4V-8m52-r6LK-ZYd3-mMHA-FtuyXV LV Write Access read/write LV Status available open 0 LV Size 1,00 GiB Current LE 256 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 256 Block device 252:7 --- Logical volume --- LV Name /dev/cinder-volumes/volume-954d2f1b-837b-4ba5-abfd-b3610597be5e VG Name cinder-volumes LV UUID belquE-WxQ2-gt6Y-WlPE-Hmq3-B9Am-zcYD3P LV Write Access read/write LV Status available open 0 LV Size 60,00 GiB Current LE 15360 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 256 Block device 252:8 --- Logical volume --- LV Name /dev/cinder-volumes/volume-05d037d1-4e61-4419-929a-fe340e00e1af VG Name cinder-volumes LV UUID Pt61e7-l3Nu-1IdX-T2sb-0GQD-PhS6-XtIIUj LV Write Access read/write LV Status available open 1 LV Size 1,00 GiB Current LE 256 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 256 Block device 252:9 --- Logical volume --- LV Name /dev/cinder-volumes/volume-316f77c6-bf13-4ea4-9b98-028198f3922f VG Name cinder-volumes LV UUID e46mBx-CRps-HYKk-aJsc-XFRd-B1Rv-UVk8gT LV Write Access read/write LV Status available open 1 LV Size 60,00 GiB Current LE 15360 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 256 Block device 252:10 Do you know how I can remove all of them? When I try lvremove /dev/cinder-volumes/volume-316f77c6-bf13-4ea4-9b98-028198f3922f, for example, I get: Can't remove open logical volume "volume-316f77c6-bf13-4ea4-9b98-028198f3922f" Thank you again. 2013/11/8 Guilherme Russi [email protected]<mailto:[email protected]> Very thanks again. Best regards. 2013/11/8 Razique Mahroua [email protected]<mailto:[email protected]> Oh yah true! not sure “conductors” exist yet for Cinder, meaning meanwhile, every node needs a direct access to the database glad to hear it’s working :) On 08 Nov 2013, at 08:53, Guilherme Russi [email protected]<mailto:[email protected]> wrote: Hello again Razique, I've found the problem, I need to add the grants on the mysql to my another IP. Now it's working really good :D I've found this link too if someone needs: http://docs.openstack.org/admin-guide-cloud/content//managing-volumes.html<http://docs.openstack.org/admin-guide-cloud/content/managing-volumes.html> Thank you so much, and if you need me just let me know. Best regards. Guilherme. 2013/11/8 Guilherme Russi [email protected]<mailto:[email protected]> Hello Razique, I got a couple of doubts, do you know if I need to do something else that's is not on the link you sent me? I'm asking because I followed the configuration but it's not working, here is what I get: I've installed the cinder-volume at the second computer that have the HD, and I've changed it's cinder.conf. I've changed too the master's cinder.conf like is following: [DEFAULT] rootwrap *config = /etc/cinder/rootwrap.conf sql*connection = mysql://cinder:password@localhost/cinder api*paste*confg = /etc/cinder/api-paste.ini iscsi_helper=iscsiadm iscsi_helper = ietadm iscsi *helper = tgtadm volume*name_template = volume-%s volume_group = cinder-volumes verbose = True auth *strategy = keystone iscsi*ip *address = 192.168.3.1 scheduler* driver=cinder.scheduler.filter_scheduler.FilterScheduler Rabbit authorization rabbit *host = localhost rabbit*port = 5672 rabbit*hosts = $rabbit*host:$rabbit *port rabbit*use *ssl = false rabbit*userid = guest rabbit_password = password rabbit*virtual*host = /nova state *path = /var/lib/cinder lock*path = /var/lock/cinder volumes_dir = /var/lib/cinder/volumes rpc*backend = cinder.rpc.impl*kombu enabled *backends=orion-1,orion-4 [orion-1] volume*group=cinder-volumes volume *driver=cinder.volume.drivers.lvm.LVMISCSIDriver volume*backend*name=LVM* iSCSI [orion-4] volume *group=cinder-volumes-2 volume* driver=cinder.volume.drivers.lvm.LVMISCSIDriver volume*backend*name=LVM_iSCSI The cinder.conf on the second computer is like this but the IPs are changed with the controller IP (It has the cinder-api), and when I run service cinder-volume restart at the second computer it's status is stop/waiting. Any ideas? Thanks :) 2013/11/8 Razique Mahroua [email protected]<mailto:[email protected]> sure :) On 08 Nov 2013, at 05:39, Guilherme Russi [email protected]<mailto:[email protected]> wrote: Oh great! I'll try here and send you the results. Very thanks :) 2013/11/8 Razique Mahroua [email protected]<mailto:[email protected]> If I’m not mistaken, you only need to install the “cinder-volume’ service that will update its status to your main node :) On 08 Nov 2013, at 05:34, Guilherme Russi [email protected]<mailto:[email protected]> wrote: Great! I was reading the link and I have one question, do I need to install cinder at the other computer too? Thanks :) 2013/11/8 Razique Mahroua [email protected]<mailto:[email protected]> Ok in that case, with Grizzly you can use the “multi-backends” feature: https://wiki.openstack.org/wiki/Cinder-multi-backend and that should do it :) On 08 Nov 2013, at 05:29, Guilherme Russi [email protected]<mailto:[email protected]> wrote: It is a hard disk, my scenario is one Controller (where I have my storage cinder and my network quantum) and four compute nodes. 2013/11/8 Razique Mahroua [email protected]<mailto:[email protected]> ok ! what is your actual Cinder backend? Is it a hard disk, a SAN, a network volume, etc… On 08 Nov 2013, at 05:20, Guilherme Russi < [email protected]<mailto:[email protected]>> wrote: Hi Razique, thank you for answering, I want to expand my cinder storage, is it the block storage? I'll use the storage to allow VMs to have more hard disk space. Regards. Guilherme. 2013/11/8 Razique Mahroua [email protected]<mailto:[email protected]> Hi Guilherme ! Which storage do you precisely want to expand? Regards, Razique On 08 Nov 2013, at 04:52, Guilherme Russi < [email protected]<mailto:[email protected]>> wrote: Hello guys, I have a Grizzly deployment running fine with 5 nodes, and I want to add more storage on it. My question is, can I install a new HD on another computer thats not the controller and link this HD with my cinder that it can be a storage too? The computer I will install my new HD is at the same network as my cloud is. I'm asking because I haven't seen a question like that here. Does anybody knows how to do that? Have a clue? Any help is welcome. Thank you all. Best regards. ------------------------------ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : [email protected]<mailto:[email protected]> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
_______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : [email protected] Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
