Ping,
Port 5672 is attached only to 172.222.12.3 and would not respond on
³localhost"
Vedu
On 8/3/15, 9:32 AM, "Ping Song" <[email protected]> wrote:
>Folks:
>
>It looks I ran into a similar issue...
>
>When I tried to launch instance I got:
>Error: Failed to launch instance "instance1": Please try again later
>[Error: No valid host was found. ].
>
>While I monitor the log I have:
>
>root@contrail:~# tail -f /var/log/cinder/cinder-scheduler.log
>2015-08-02 23:47:52.088 2037 ERROR oslo.messaging._drivers.impl_rabbit
>[req-3d2dd035-c205-4206-8e3f-7ad9e2c1430a - - - - -] AMQP server on
>localhost:5672 is unreachable: [Errno 111] ECONNREFUSED. Trying again in
>30 seconds.
>2015-08-02 23:48:22.114 2037 INFO oslo.messaging._drivers.impl_rabbit
>[req-3d2dd035-c205-4206-8e3f-7ad9e2c1430a - - - - -] Delaying reconnect
>for 1.0 seconds...
>2015-08-02 23:48:23.116 2037 INFO oslo.messaging._drivers.impl_rabbit
>[req-3d2dd035-c205-4206-8e3f-7ad9e2c1430a - - - - -] Connecting to AMQP
>server on localhost:5672
>2015-08-02 23:48:23.131 2037 ERROR oslo.messaging._drivers.impl_rabbit
>[req-3d2dd035-c205-4206-8e3f-7ad9e2c1430a - - - - -] AMQP server on
>localhost:5672 is unreachable: [Errno 111] ECONNREFUSED. Trying again in
>30 seconds.
>
>I'm not sure if this is related...but
>When I look at the status of the service I don't see anything goes
>wrong...
>
>root@contrail:~# rabbitmqctl status
>Status of node 'rabbit@contrail-ctrl' ...
>[{pid,11474},
> {running_applications,[{rabbit,"RabbitMQ","3.5.0"},
> {os_mon,"CPO CXC 138 46","2.2.14"},
> {mnesia,"MNESIA CXC 138 12","4.11"},
> {xmerl,"XML parser","1.3.5"},
> {sasl,"SASL CXC 138 11","2.3.4"},
> {stdlib,"ERTS CXC 138 10","1.19.4"},
> {kernel,"ERTS CXC 138 10","2.16.4"}]},
> {os,{unix,linux}},
> {erlang_version,"Erlang R16B03 (erts-5.10.4) [source] [64-bit] [smp:8:8]
>[async-threads:30] [kernel-poll:true]\n"},
> {memory,[{total,38403376},
> {connection_readers,437080},
> {connection_writers,0},
> {connection_channels,0},
> {connection_other,426992},
> {queue_procs,2704},
> {queue_slave_procs,0},
> {plugins,0},
> {other_proc,13738216},
> {mnesia,60176},
> {mgmt_db,0},
> {msg_index,46784},
> {other_ets,785752},
> {binary,400560},
> {code,16351158},
> {atom,561761},
> {other_system,5592193}]},
> {alarms,[]},
> {listeners,[{clustering,25672,"::"},{amqp,5672,"172.222.12.3"}]},
> {vm_memory_high_watermark,0.4},
> {vm_memory_limit,3269242060},
> {disk_free_limit,50000000},
> {disk_free,7446581248},
> {file_descriptors,[{total_limit,10140},
> {total_used,52},
> {sockets_limit,9124},
> {sockets_used,50}]},
> {processes,[{limit,1048576},{used,278}]},
> {run_queue,0},
>
>root@contrail:~# netstat -lntp | grep 5672
>tcp 0 0 172.222.12.3:5672 0.0.0.0:*
>LISTEN 11474/beam.smp
>tcp 0 0 0.0.0.0:25672 0.0.0.0:*
>LISTEN 11474/beam.smp
>
>root@contrail:~# service rabbitmq-server status
>rabbitmq-server RUNNING pid 11466, uptime 0:07:19
>
>so 172.222.12.3 is my data interface (eth1) , while my management
>interface is eth0...
>
>any advice?
>
>Regards
>ping
>
>-----Original Message-----
>From: Lluís Gifre [mailto:[email protected]]
>Sent: Monday, May 11, 2015 6:53 AM
>To: EXT - [email protected] <[email protected]>; Saravanan
>Purushothaman <[email protected]>; ask-contrail
><[email protected]>; [email protected]
>Cc: contrailuser <[email protected]>
>Subject: Re: [Users] Error creating volumes with cinder
>
>Hi all,
>
>Regarding this problem, I tried something that works for me:
>
>I noticed that using Contrail v2.10 I cannot install the cinder-volume
>service because this package is not available, so I decided to create my
>own package.
>
>These are the steps I followed in Ubuntu 14.04.1:
>
>I'm using LVM with a volume group named cinder-volumes to store the
>volumes.
>
>
># 1. Create my cinder-volume_2014.1.3-0ubuntu1~cloud0_all.deb package #
>1.1. download original cinder package
>https://launchpadlibrarian.net/187842740/cinder-volume_2014.1.3-0ubuntu1.1
>_all.deb
>
># 1.2. create temporal folder
>mkdir mycinder
>cd mycinder
>
># 1.3. extract package files
>dpkg-deb -x ../cinder-volume_2014.1.3-0ubuntu1.1_all.deb .
>
># 1.4. extract package control files
>dpkg-deb -e ../cinder-volume_2014.1.3-0ubuntu1.1_all.deb
>
># 1.5. edit file DEBIAN/control and change package version and dependency
>in cinder-common # replace "2014.1.3-0ubuntu1.1" by
>"1:2014.1.3-0ubuntu1~cloud0" so next lines should result
> Version: 1:2014.1.3-0ubuntu1~cloud0
> Depends: cinder-common (= 1:2014.1.3-0ubuntu1~cloud0), lvm2, tgt,
>sysv-rc (>= 2.88dsf-24) | file-rc (>= 0.8.16), python:any
> Keep the rest of file as is
>
># 1.6. generate new package
>dpkg-deb -b . ../cinder-volume_2014.1.3-0ubuntu1~cloud0_all.deb
>dpkg-deb -i . ../cinder-volume_2014.1.3-0ubuntu1~cloud0_all.deb
>
># 2. Configure database in openstack node where cinder api/scheduler is
>installed:
># 2.1. Take root password used for the database creation grep "mysql -u
>root --password=" /opt/contrail/utils/setup_*
>
># 2.2. Login into database and grant permissions to cinder user mysql -u
>root --password=<password taken from previous section>
>mysql> GRANT ALL PRIVILEGES ON cinder.* TO 'cinder'@'localhost'
>IDENTIFIED BY 'cinder';
>mysql> GRANT ALL PRIVILEGES ON cinder.* TO 'cinder'@'<your openstack
>node hostname>' IDENTIFIED BY 'cinder';
>mysql> GRANT ALL PRIVILEGES ON cinder.* TO 'cinder'@'%' IDENTIFIED BY
>'cinder';
>mysql> exit
>
># 2.3. Populate cinder database
>su -s /bin/sh -c "cinder-manage db sync" cinder
>
># 3. Install package
># 3.1. Copy new package cinder-volume_2014.1.3-0ubuntu1~cloud0_all.deb
>into compute nodes
>
># 3.2. Configure cinder api/scheduler
>nano /etc/cinder/cinder.conf
> [DEFAULT]
> rootwrap_config = /etc/cinder/rootwrap.conf
> api_paste_config = /etc/cinder/api-paste.ini
> iscsi_helper = tgtadm
> volume_name_template = volume-%s
> volume_group = cinder-volumes
> verbose = True
> auth_strategy = keystone
> state_path = /var/lib/cinder
> lock_path = /var/lock/cinder
> volumes_dir = /var/lib/cinder/volumes
>
> rpc_backend = rabbit
> rabbit_host = <your_openstack_node_IP>
> rabbit_port = 5672
> my_ip = <your_openstack_node_IP>
> glance_host = <your_openstack_node_IP>
>
> osapi_volume_workers = 4
>
> [database]
> connection =
>mysql://cinder:cinder@<your_cinder_api/scheduler_node_IP>/cinder
>
> [keystone_authtoken]
> admin_tenant_name = service
> admin_user = cinder
> admin_password = (copy the value from same attribute in
>/etc/nova/nova.conf)
> auth_protocol = http
> auth_host = <your_openstack_node_IP>
> auth_port = 35357
> auth_uri = http://<your_openstack_node_IP>:5000
>
># 3.3. Restart cinder-api and cinder-scheduler services in openstack node
>service cinder-api restart service cinder-scheduler restart
>
># 4. Install new cinder-volume in compute nodes # 4.1. copy
>cinder-volumes package into contrail install repo cp
>cinder-volume_2014.1.3-0ubuntu1~cloud0_all.deb
>/opt/contrail/contrail_install_repo
>
># 4.2. move to contrail install repo folder cd
>/opt/contrail/contrail_install_repo
>
># 4.3. regenerate Packages.gz
>dpkg-scanpackages . /dev/null | gzip -9c > Packages.gz
>
># 4.4. update apt-get cache and install cinder-volume apt-get update
>apt-get install cinder-volume
>
># 5. Configure compute node
># 5.1. Configure cinder-volume service
>nano /etc/cinder/cinder.conf
> [DEFAULT]
> rootwrap_config = /etc/cinder/rootwrap.conf
> api_paste_config = /etc/cinder/api-paste.ini
> iscsi_helper = tgtadm
> volume_name_template = volume-%s
> volume_group = cinder-volumes
> #verbose = True
> auth_strategy = keystone
> state_path = /var/lib/cinder
> lock_path = /var/lock/cinder
> volumes_dir = /var/lib/cinder/volumes
>
> rpc_backend = rabbit
> rabbit_host = <your_openstack_node_IP>
> rabbit_port = 5672
> my_ip = <your_compute_node_IP>
> glance_host = <your_openstack_node_IP>
>
> [database]
> connection =
>mysql://cinder:cinder@<your_openstack_node_IP>/cinder
>
> [keystone_authtoken]
> admin_tenant_name = service
> admin_user = cinder
> admin_password = (copy the value from same attribute in
>/etc/nova/nova.conf)
> auth_protocol = http
> auth_host = <your_openstack_node_IP>
> auth_port = 35357
> auth_uri = http://<your_openstack_node_IP>:5000
>
># 5.2. Configure LVM service
>nano /etc/lvm/lvm.conf
> Leave file as is, except for the filter attribute in devices
>section.
> The device specified (sda4) has to be the same where
>cinder-volumes resides.
>
> devices {
> ...
> # Cinder filter
> filter = [ "a/sda4/", "r/.*/" ]
> ...
> }
>
># 5.3. Restart cinder-volume and tgt services service cinder-volume
>restart service tgt restart
>
>Right now you should be able to create instances specifying to create a
>volume from an image or create empty volumes without problem.
>
>The only issue I'm experiencing is that Openstack's Dashboard, in Admin
>> Hypervisors reports an incorrect storage total space and used space.
>Except for this, it works correctly.
>
>Hope it helps.
>Lluis
>
>
>
>El 14/04/15 a les 18:04, Joao Prino ha escrit:
>> Hi Saravanan,
>>
>> Thanks for the update.
>> The output:
>>
>> # dpkg -l| grep cinder
>> ii cinder-api 1:2014.1.3-0ubuntu1~cloud0
>> all Cinder storage service - API server
>> ii cinder-common 1:2014.1.3-0ubuntu1~cloud0
>> all Cinder storage service - common files
>> ii cinder-scheduler 1:2014.1.3-0ubuntu1~cloud0
>> all Cinder storage service - Scheduler server
>> ii python-cinder 1:2014.1.3-0ubuntu1~cloud0
>> all Cinder Python libraries
>> ii python-cinderclient 1:1.0.8-0ubuntu1
>> all python bindings to the OpenStack Volume API
>>
>> That might explain the behaviour, let me install the packages and will
>>get back to you.
>>
>>
>> Cheers,
>> Joao
>>
>> -----Original Message-----
>> From: Saravanan Purushothaman [mailto:[email protected]]
>> Sent: 14 April 2015 16:56
>> To: Joao Prino; Lluís Gifre; ask-contrail; [email protected]
>> Cc: contrailuser
>> Subject: RE: [Users] Error creating volumes with cinder
>>
>>
>> Hi Joao,
>>
>> Can you try this.. "dpkg -l| grep cinder"
>>
>> I think cinder-volume package is not installed.
>>
>> if cinder-volume service is not running then we may get
>>this "No valid host was found"
>>
>> NOTE:
>> cinder-volume is storage related package.
>> If you want storage on contrail then please follow the link
>> instructions to install contrail-storage package. (cinder-volume will
>> be installed as part of contrail-storage package)
>>
>> you can get the "contrail storage packages"
>> from this link
>> http://www.juniper.net/support/downloads/?p=contrail#sw
>>
>>
>> Regards,
>> Saravanan
>>
>> -----Original Message-----
>> From: Joao Prino [mailto:[email protected]]
>> Sent: Tuesday, April 14, 2015 12:37 AM
>> To: Lluís Gifre; ask-contrail; [email protected]
>> Cc: contrailuser
>> Subject: RE: [Users] Error creating volumes with cinder
>>
>> Hello Lluis Dev,
>>
>> I'm calling Dev's attention to this problem.
>> Look forward to hear back whenever possible.
>>
>> Cheers,
>> Joao
>>
>> -----Original Message-----
>> From: Lluís Gifre [mailto:[email protected]]
>> Sent: 10 April 2015 17:03
>> To: Joao Prino; Adrian Smith
>> Cc: contrailuser
>> Subject: Re: [Users] Error creating volumes with cinder
>>
>> Hi Joao,
>>
>> Yes, I have already done this correciton but the problem "No backend
>>provided" and "No valid host was found" while the cinder-scheduler is
>>creating the volume is still happening.
>>
>> Thanks,
>> Lluis
>>
>>
>> El 10/04/15 a les 17:54, Joao Prino ha escrit:
>>> Hi Lluis,
>>>
>>> I had this exact same problem where was able to solve it by adding
>>>the "rabbit_host = <ip>" under '/etc/cinder/cinder.conf', editing
>>>'/etc/rabbitmq/rabbitmq.config' to match my cluster setup's fqdn
>>>(wrongly populated) and changed the existing "rabbit_host = <ip>" from
>>>'/etc/nova/nova.conf as it had the wrong IP assigned to it.
>>> This problem was due to having two NICs where RabbitMQ was given my
>>>NIC1 address settings and it should have had my NIC2's address
>>>according to the testbed file..
>>>
>>> Hope it helps!
>>>
>>> Joao
>>>
>>>
>>> -----Original Message-----
>>> From: Users [mailto:[email protected]] On Behalf
>>> Of Lluís Gifre
>>> Sent: 10 April 2015 15:19
>>> To: Adrian Smith
>>> Cc: contrailuser
>>> Subject: Re: [Users] Error creating volumes with cinder
>>>
>>> Hi Adrian,
>>>
>>> Thanks for your quick answer.
>>> I checked the log you mentioned but I hadn't found any interesting
>>>thing.
>>>
>>> Do you have any other idea?
>>>
>>> Thanks,
>>> Lluis
>>>
>>> El 10/04/15 a les 13:14, Adrian Smith ha escrit:
>>>> Hi Lluis,
>>>>
>>>> Take a look in the cinder scheduler log,
>>>> /var/log/cinder/scheduler.log. It should have a more meaningful error.
>>>>
>>>> Adrian
>>>>
>>>> On 10 April 2015 at 11:48, Lluís Gifre <[email protected]> wrote:
>>>>> Dear all,
>>>>>
>>>>> I'm deploying a opencontrail 2.10 + openstack icehouse testbed in
>>>>> Ubuntu 14.04.
>>>>> I tried it in many ways (with physical and virtual machines, single
>>>>> and multi-box testbeds).
>>>>>
>>>>> Right now I'm trying with a single virtual machine with LVM and the
>>>>> cinder-volumes volume group created.
>>>>>
>>>>> The deploy completes successfully.
>>>>>
>>>>> However, just after completing the set-up, I realized that in
>>>>> cinder-schedule.log the next message is added periodically:
>>>>>
>>>>> 2015-04-10 07:27:29.426 2315 INFO
>>>>> oslo.messaging._drivers.impl_rabbit
>>>>> [req-e77c84ab-49f6-40f2-bb20-1c62b23605d6 - - - - -] Reconnecting
>>>>> to AMQP server on localhost:5672
>>>>> 2015-04-10 07:27:29.427 2315 INFO
>>>>> oslo.messaging._drivers.impl_rabbit
>>>>> [req-e77c84ab-49f6-40f2-bb20-1c62b23605d6 - - - - -] Delaying
>>>>> reconnect for
>>>>> 1.0 seconds...
>>>>> 2015-04-10 07:27:30.442 2315 ERROR
>>>>> oslo.messaging._drivers.impl_rabbit
>>>>> [req-e77c84ab-49f6-40f2-bb20-1c62b23605d6 - - - - -] AMQP server on
>>>>> localhost:5672 is unreachable: [Errno 111] ECONNREFUSED. Trying
>>>>> again in 30 seconds.
>>>>>
>>>>> To solve the problem I edited /etc/cinder/cinder.conf and added:
>>>>>
>>>>> rabbit_host = 192.168.67.13 # IP address taken from
>>>>> /etc/rabbitmq/rabbitmq.config rabbit_port = 5672
>>>>>
>>>>> Content of my /etc/rabbitmq/rabbitmq.config:
>>>>>
>>>>> [
>>>>> {rabbit, [ {tcp_listeners, [{"192.168.67.13", 5672}]},
>>>>> {loopback_users, []},
>>>>> {log_levels,[{connection, info},{mirroring, info}]} ]
>>>>> }
>>>>> ].
>>>>>
>>>>> After rebooting, this change solved the recurrent reconnecting to
>>>>> AMQP problem.
>>>>> Right now the cinder-scheduler log shows:
>>>>>
>>>>> 2015-04-10 07:32:44.063 2525 AUDIT cinder.service [-] Starting
>>>>> cinder-scheduler node (version 2014.1.3)
>>>>> 2015-04-10 07:32:44.197 2525 INFO
>>>>> oslo.messaging._drivers.impl_rabbit
>>>>> [req-21e86bf9-3fa6-4ab0-9968-616242200c41 - - - - -] Connected to
>>>>> AMQP server on 192.168.67.13:5672
>>>>> 2015-04-10 07:32:48.383 2525 INFO
>>>>> oslo.messaging._drivers.impl_rabbit
>>>>> [-] Connected to AMQP server on 192.168.67.13:5672
>>>>>
>>>>> Next I tried to create a volume, no mater if I do it from the
>>>>> openstack dashboard or using cinder's or nova's command line
>>>>> interface, the result is an error. For example trying with cinder
>>>>>command line:
>>>>>
>>>>> # cinder list
>>>>>
>>>>>+----+--------+--------------+------+-------------+----------+--------
>>>>>-----+
>>>>> | ID | Status | Display Name | Size | Volume Type | Bootable |
>>>>> | Attached to |
>>>>>
>>>>>+----+--------+--------------+------+-------------+----------+--------
>>>>>-----+
>>>>>
>>>>>+----+--------+--------------+------+-------------+----------+--------
>>>>>-----+
>>>>>
>>>>> # cinder create --display-name test 1
>>>>> +---------------------+--------------------------------------+
>>>>> | Property | Value |
>>>>> +---------------------+--------------------------------------+
>>>>> | attachments | [] |
>>>>> | availability_zone | nova |
>>>>> | bootable | false |
>>>>> | created_at | 2015-04-10T07:57:12.787493 |
>>>>> | display_description | None |
>>>>> | display_name | test |
>>>>> | encrypted | False |
>>>>> | id | d374b571-5df1-47f3-ae6e-c3218aebb9db |
>>>>> | metadata | {} |
>>>>> | size | 1 |
>>>>> | snapshot_id | None |
>>>>> | source_volid | None |
>>>>> | status | creating |
>>>>> | volume_type | None |
>>>>> +---------------------+--------------------------------------+
>>>>>
>>>>> # cinder list
>>>>>
>>>>>+--------------------------------------+--------+--------------+------
>>>>>+-------------+----------+-------------+
>>>>> | ID | Status | Display Name |
>>>>>Size |
>>>>> Volume Type | Bootable | Attached to |
>>>>>
>>>>>+--------------------------------------+--------+--------------+------
>>>>>+-------------+----------+-------------+
>>>>> | d374b571-5df1-47f3-ae6e-c3218aebb9db | error | test | 1
>>>>>|
>>>>> None | false | |
>>>>>
>>>>>+--------------------------------------+--------+--------------+------
>>>>>+-------------+----------+-------------+
>>>>>
>>>>> Then I checked the /var/log/cinder/cinder-scheduler.log and found
>>>>> next
>>>>> message:
>>>>>
>>>>> 2015-04-10 07:36:14.547 2525 WARNING cinder.context [-] Arguments
>>>>> dropped when creating context: {'user':
>>>>> u'ad2be0cd57c44177b1ba9b0735ea0f44',
>>>>> 'tenant': u'90aab64e7e2b45f3891b0f08978e063e', 'user_identity':
>>>>> u'ad2be0cd57c44177b1ba9b0735ea0f44 90aab64e7e2b45f3891b0f08978e063e
>>>>> -
>>>>> - -'}
>>>>> 2015-04-10 07:36:14.629 2525 ERROR
>>>>> cinder.scheduler.flows.create_volume
>>>>> [req-ba5d16aa-5092-4f5f-a4c6-d9481df83bfd
>>>>> ad2be0cd57c44177b1ba9b0735ea0f44 90aab64e7e2b45f3891b0f08978e063e -
>>>>> - -] Failed to schedule_create_volume: No valid host was found.
>>>>>
>>>>> I have enabled DEBUG by adding debug = True to
>>>>> /etc/cinder/cinder.conf and found in the cinder-scheduler.log: ...
>>>>>No backend provided ...
>>>>> I installed and condifured tgtd according to
>>>>> http://rconradharris.com/2013/01/14/getting-cinder-up-and-running.h
>>>>> t
>>>>> m
>>>>> l
>>>>>
>>>>> But the error still remains.
>>>>>
>>>>> Do somebody have an idea of how to solve this problem?
>>>>>
>>>>> Thanks,
>>>>> Lluis
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list
>>>>> [email protected]
>>>>> http://lists.opencontrail.org/mailman/listinfo/users_lists.opencont
>>>>> r
>>>>> a
>>>>> il.org
>>> _______________________________________________
>>> Users mailing list
>>> [email protected]
>>> http://lists.opencontrail.org/mailman/listinfo/users_lists.opencontra
>>> i
>>> l.org
>
_______________________________________________
Users mailing list
[email protected]
http://lists.opencontrail.org/mailman/listinfo/users_lists.opencontrail.org