Dear All...

I'm facing some problems while deploying a 2nd cinder volume in a server different from the one where cinder scheduler is deployed (if this is not the right forum to ask, please redirect me to the proper support mailing list, TIA).

The volumes are perfectly deployed in the 1st cinder-volume instance (where cinder-scheduler and cinder api are running) but on my 2nd cinder-volume, the volume stays in creating state forever.

The setup is has following:
 - I'me using HAVANA
- cloud03 is the server where cinder-scheduler, cinder-api and cinder-volume is installed
 - cloud04 is the server where only cinder-volume will be installed
- cloud03 and cloud04 are configured so that all connections are accepted in the machines firewalls

Let me guide you through my configuration steps:

1) I've created a pv and vg in cloud04

# pvdisplay
  --- Physical volume ---
  PV Name               /dev/loop0
  VG Name               cinder-volumes
  PV Size               10.00 GiB / not usable 4.00 MiB
  Allocatable           yes
  PE Size               4.00 MiB
  Total PE              2559
  Free PE               2559
  Allocated PE          0
  PV UUID               RqQa2d-Fw3u-EdR4-Yr6Y-oZVv-fHVT-IkKxLT

# vgdisplay
  --- Volume group ---
  VG Name               cinder-volumes
  System ID
  Format                lvm2
  Metadata Areas        1
  Metadata Sequence No  1
  VG Access             read/write
  VG Status             resizable
  MAX LV                0
  Cur LV                0
  Open LV               0
  Max PV                0
  Cur PV                1
  Act PV                1
  VG Size               10.00 GiB
  PE Size               4.00 MiB
  Total PE              2559
  Alloc PE / Size       0 / 0
  Free  PE / Size       2559 / 10.00 GiB
  VG UUID               Uk5grr-qIWs-uQT2-1gRS-fRIa-4bd7-Ik66or

2) I've installed and configured cinder-volume. The definitions are:

# cat /etc/cinder/cinder.conf | grep -v "#"
[DEFAULT]
api_paste_config=api-paste.ini
state_path=/var/lib/cinder
rootwrap_config=/etc/cinder/rootwrap.conf
auth_strategy=keystone
volume_name_template=volume-%s
verbose=true
use_stderr=true
rpc_backend=cinder.openstack.common.rpc.impl_qpid
qpid_hostname=<cloud03 ip>
iscsi_ip_address=<cloud04 ip>
volume_clear=none
volume_group=cinder-volumes
sql_connection=mysql://cinder:xxxxxxxxxxxxxx@<cloud03 ip>/cinder
qpid_reconnect_timeout=0
qpid_reconnect_limit=0
qpid_reconnect=True
qpid_reconnect_interval_max=0
qpid_reconnect_interval_min=0
sql_idle_timeout=3600
qpid_reconnect_interval=0
notification_driver=cinder.openstack.common.notifier.rpc_notifier

# grep include /etc/tgt/targets.conf
include /etc/cinder/volumes/*


3) After restarting cinder and tgt, cinder seems to acknowledge the new node

# cinder-manage host list
host                    zone
cloud03.my.domain       nova
cloud04.my.domain       nova

# cinder-manage service list
Binary Host Zone Status State Updated At cinder-volume cloud03 nova enabled :-) 2014-03-22 10:41:12 cinder-scheduler cloud03 nova enabled :-) 2014-03-22 10:41:05 cinder-backup cloud03 nova enabled :-) 2014-03-22 10:41:05 cinder-volume cloud04 nova enabled :-) 2014-03-22 10:41:03

4) However, when I try to create a new volume (in the case where all volume space is already used in the working cloud03 machine) it never leaves the creating status,

# cinder create 6
+---------------------+--------------------------------------+
|       Property      |                Value                 |
+---------------------+--------------------------------------+
|     attachments     |                  []                  |
|  availability_zone  |                 nova                 |
|       bootable      |                false                 |
|      created_at     |      2014-03-22T11:17:50.918263      |
| display_description |                 None                 |
|     display_name    |                 None                 |
|          id         | 7ab9f0c2-63d6-499f-be18-c6b5984fece2 |
|       metadata      |                  {}                  |
|         size        |                  6                   |
|     snapshot_id     |                 None                 |
|     source_volid    |                 None                 |
|        status       |               creating               |
|     volume_type     |                 None                 |
+---------------------+--------------------------------------+
[root@cloud04 ~(keystone_admin)]# cinder list
+--------------------------------------+-----------+--------------+------+-------------+----------+-------------+
| ID | Status | Display Name | Size | Volume Type | Bootable | Attached to |
+--------------------------------------+-----------+--------------+------+-------------+----------+-------------+
| 5eb182b3-fa19-42da-bb08-df3c54df3c2a | available | 15 | 15 | None | false | | | 7ab9f0c2-63d6-499f-be18-c6b5984fece2 | creating | None | 6 | None | false | | | a2423edb-3147-4082-98ae-23dde68fa19a | available | 1 | 1 | None | false | |
+--------------------------------------+-----------+--------------+------+-------------+----------+-------------+
[root@cloud04 ~(keystone_admin)]# cinder show 7ab9f0c2-63d6-499f-be18-c6b5984fece2
+--------------------------------+--------------------------------------+
| Property | Value |
+--------------------------------+--------------------------------------+
| attachments | [] | | availability_zone | nova | | bootable | false | | created_at | 2014-03-22T11:17:50.000000 | | display_description | None | | display_name | None | | id | 7ab9f0c2-63d6-499f-be18-c6b5984fece2 | | metadata | {} | | os-vol-host-attr:host | cloud04.ncg.ingrid.pt | | os-vol-mig-status-attr:migstat | None | | os-vol-mig-status-attr:name_id | None | | os-vol-tenant-attr:tenant_id | 5fa11a8e9a9b40ea887f6d31d299ba7c | | size | 6 | | snapshot_id | None | | source_volid | None | | status | creating | | volume_type | None |
+--------------------------------+--------------------------------------+

Looking to the volume log on cloud04, I see messages announcing that the cinder-volume VG is available to allow the creation of new volumes:

2014-03-22 11:17:48.502 13634 DEBUG qpid.messaging [-] RACK[21c9290]: Message({'oslo.message': '{"_context_roles": ["admin"], "_context_request_id": "req-4641459f-92a2-4bbf-bd6b-206f60fc80ca", "_context_quota_class": null, "_context_service_catalog": [], "_context_tenant": null, "args": {"service_name": "volume", "host": "cloud04.ncg.ingrid.pt", "capabilities": {"QoS_support": false, "location_info": "LVMVolumeDriver:cloud04.ncg.ingrid.pt:cinder-volumes:default:0", "volume_backend_name": "LVM_iSCSI", "free_capacity_gb": 10.0, "driver_version": "2.0.0", "total_capacity_gb": 10.0, "reserved_percentage": 0, "vendor_name": "Open Source", "storage_protocol": "iSCSI"}}, "_unique_id": "2ae40e9c8ab74a33bf3287d219167707", "_context_timestamp": "2014-03-22T11:17:48.492982", "_context_user_id": null, "_context_project_name": null, "_context_read_deleted": "no", "_context_auth_token": null, "namespace": null, "_context_is_admin": true, "version": "1.0", "_context_project_id": null, "_context_user": null, "method": "update_service_capabilities", "_context_remote_address": null}', 'oslo.version': '2.0'}) msg_acked /usr/lib/python2.6/site-packages/qpid/messaging/driver.py:1269

but I do not see messages regarding the specific volume creation request we just issued.


So, any help is appreciated at this point

TIA

Cheers
Goncalo



_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to     : [email protected]
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack

Reply via email to