Could you also paste the output of "openstack compute service list" and "openstack network agent list"? I'd like to see if the nova and neutron services are all up and running.

Note that I don't have cinder installed. The docs said the block storage
service was optional.

You're right, I just assumed that it's also installed.

the config files should be the same as the site.

I can't see any obvious error there, although it differs from our configs since we run Ocata, of course.

I don't think the
warning about the  placement api is relevant. What about the other one:
Unable to refresh my resource provider record?

I'm not sure about that, it is just a warning.
Can you confirm that glance is working properly and the image is okay? Is the network layout as expected? Any information in other logs like neutron and glance?

Eugen


Zitat von Sashan Govender <sash...@gmail.com>:

Note that I don't have cinder installed. The docs said the block storage
service was optional. I was following the newton guide here:
https://docs.openstack.org/newton/install-guide-rdo/launch-instance-provider.html
so
the config files should be the same as the site. Running a controller and
compute node using centos 7 in kvm with nested kvm turned on on my host
machine. Firewall is disabled on controller and compute node.

Contents of nova.conf on the compute node:

[DEFAULT]
enabled_apis = osapi_compute,metadata
transport_url = rabbit://openstack:rootroot@controller
auth_strategy = keystone
my_ip = 192.168.122.5
use_neutron = True
firewall_driver = nova.virt.firewall.NoopFirewallDriver


[keystone_authtoken]
auth_uri = http://controller:5000
auth_url = http://controller:35357
memcached_servers = controller:11211
auth_type = password
project_domain_name = Default
user_domain_name = Default
project_name = service
username = nova
password = rootroot

[vnc]
enabled = True
vncserver_listen = 0.0.0.0
vncserver_proxyclient_address = $my_ip
novncproxy_base_url = http://controller:6080/vnc_auto.html

[glance]
api_servers = http://controller:9292

[oslo_concurrency]
lock_path = /var/lib/nova/tmp

Contents from nova.conf on the controller:

[DEFAULT]
auth_strategy=keystone
my_ip=192.168.122.186
use_neutron=true
enabled_apis=osapi_compute,metadata
firewall_driver=nova.virt.firewall.NoopFirewallDriver
debug=false
transport_url=rabbit://openstack:rootroot@controller

[api_database]
connection=mysql+pymysql://nova:rootroot@controller/nova_api

[database]
connection=mysql+pymysql://nova:rootroot@controller/nova
[glance]
api_servers=http://controller:9292

[keystone_authtoken]
auth_uri = http://controller:5000
auth_url = http://controller:35357
memcached_servers = controller:11211
auth_type = password
project_domain_name = Default
user_domain_name = Default
project_name = service
username = nova
password = rootroot

[neutron]
url = http://controller:9696
auth_url = http://controller:35357
auth_type = password
project_domain_name = Default
user_domain_name = Default
region_name = RegionOne
project_name = service
username = neutron
password = rootroot
service_metadata_proxy = True
metadata_proxy_shared_secret = sharedsecret

[vnc]
vncserver_listen=$my_ip
vncserver_proxyclient_address=$my_ip

Output from various commands:

[sashan@controller ~]$  openstack flavor list
+----+---------+-----+------+-----------+-------+-----------+
| ID | Name    | RAM | Disk | Ephemeral | VCPUs | Is Public |
+----+---------+-----+------+-----------+-------+-----------+
| 0  | m1.nano |  64 |    1 |         0 |     1 | True      |
+----+---------+-----+------+-----------+-------+-----------+
[sashan@controller ~]$

[sashan@controller ~]$ openstack service list
+----------------------------------+----------+----------+
| ID                               | Name     | Type     |
+----------------------------------+----------+----------+
| 225d12b19b9d47e890537acbfa25d1ed | nova     | compute  |
| 84913aec768d4734a912c62965ba0462 | keystone | identity |
| ee5afcaa96c64baba15e7fa9cf02672f | glance   | image    |
| f862d1a46c474cb284ef381525948b8d | neutron  | network  |
+----------------------------------+----------+----------+
[sashan@controller ~]$

 openstack server list
+--------------------------------------+-------------------+--------+----------+------------+
| ID                                   | Name              | Status |
Networks | Image Name |
+--------------------------------------+-------------------+--------+----------+------------+
| 105b46c4-d30e-4a7c-99fd-ea8cbc8b43c3 | provider-instance | ERROR  |
    | cirros     |
| 503b7a3a-66c2-4de7-8106-0958e595771f | provider-instance | ERROR  |
    | cirros     |
| 62e2ec41-8711-4470-881d-d63bd5b348c0 | provider-instance | ERROR  |
    | cirros     |
| 6c0fe163-d587-41a8-89c7-18085e80bd38 | provider-instance | ERROR  |
    | cirros     |
| 2849a455-b5a4-4dd5-a12e-6fd2497eed9e | provider-instance | ERROR  |
    | cirros     |
| 465acda2-0d1a-47e3-9e48-2e63e7cf7a30 | provider-instance | ERROR  |
    | cirros     |
+--------------------------------------+-------------------+--------+----------+------------+
[sashan@controller ~]$


Content from nova-compute.log on the compute node. I don't think the
warning about the  placement api is relevant. What about the other one:
Unable to refresh my resource provider record?

[root@compute ~]# tail /var/log/nova/nova-compute.log
2018-01-16 11:16:32.209 1435 INFO nova.compute.resource_tracker
[req-08a524f9-4c31-45b9-808a-9cd4050b0e00 - - - - -] Final resource view:
name=compute phys_ram=2047MB used_ram=512MB phys_disk=16GB used_disk=0GB
total_vcpus=2 used_vcpus=0 pci_stats=[]
2018-01-16 11:16:32.236 1435 WARNING nova.scheduler.client.report
[req-08a524f9-4c31-45b9-808a-9cd4050b0e00 - - - - -] Unable to refresh my
resource provider record
2018-01-16 11:16:32.236 1435 INFO nova.compute.resource_tracker
[req-08a524f9-4c31-45b9-808a-9cd4050b0e00 - - - - -] Compute_service record
updated for compute:compute
2018-01-16 11:17:33.076 1435 INFO nova.compute.resource_tracker
[req-08a524f9-4c31-45b9-808a-9cd4050b0e00 - - - - -] Auditing locally
available compute resources for node compute
2018-01-16 11:17:33.129 1435 WARNING nova.scheduler.client.report
[req-08a524f9-4c31-45b9-808a-9cd4050b0e00 - - - - -] No authentication
information found for placement API. Placement is optional in Newton, but
required in Ocata. Please enable the placement service before upgrading.
2018-01-16 11:17:33.130 1435 WARNING nova.scheduler.client.report
[req-08a524f9-4c31-45b9-808a-9cd4050b0e00 - - - - -] Unable to refresh my
resource provider record
2018-01-16 11:17:33.168 1435 INFO nova.compute.resource_tracker
[req-08a524f9-4c31-45b9-808a-9cd4050b0e00 - - - - -] Total usable vcpus: 2,
total allocated vcpus: 0
2018-01-16 11:17:33.168 1435 INFO nova.compute.resource_tracker
[req-08a524f9-4c31-45b9-808a-9cd4050b0e00 - - - - -] Final resource view:
name=compute phys_ram=2047MB used_ram=512MB phys_disk=16GB used_disk=0GB
total_vcpus=2 used_vcpus=0 pci_stats=[]
2018-01-16 11:17:33.197 1435 WARNING nova.scheduler.client.report
[req-08a524f9-4c31-45b9-808a-9cd4050b0e00 - - - - -] Unable to refresh my
resource provider record
2018-01-16 11:17:33.197 1435 INFO nova.compute.resource_tracker
[req-08a524f9-4c31-45b9-808a-9cd4050b0e00 - - - - -] Compute_service record
updated for compute:compute


On Mon, Jan 15, 2018 at 8:05 PM Eugen Block <ebl...@nde.ag> wrote:

Hi,

you should check your config settings again, especially the "auth_url"
settings in the section(s) "[keystone_authtoken]" of all the config
files.
Are all the services up (nova, cinder and neutron) and running? What
is the output of 'nova service-list'?
Have you checked other log files for errors? Is there something
interesting in nova-compute.log?

Regards,
Eugen


Zitat von Sashan Govender <sash...@gmail.com>:

> Hi
>
> I've setup an openstack system based on the instructions here:
>
> https://docs.openstack.org/newton/
>
> I'm trying to launch an instance:
> $ . demo-openrc
> $ openstack server create --flavor m1.nano --image cirros --nic
> net-id=da77f469-f594-42f6-ab18-8b907b3359e4 --security-group default
> --key-name mykey provider-instance
>
> but get this error in the nova-conductor log file:
>
> 2018-01-15 15:46:48.938 2566 WARNING nova.scheduler.utils
> [req-5b47171a-f74e-4e8e-8659-89cce144f284
82858c289ca444bf90fcd41123d069ce
> 61b0b2b23b08419596bd923f2c544956 - - -] [instance:
> e1cfc9a9-9c21-435f-a9dc-c7c692e06c29] Setting instance to ERROR state.
> 2018-01-15 16:09:51.026 2567 ERROR nova.scheduler.utils
> [req-afff24dc-1ee0-469f-9d99-2abcb4810c7a
82858c289ca444bf90fcd41123d069ce
> 61b0b2b23b08419596bd923f2c544956 - - -] [instance:
> 0ba01247-5513-4c58-bf04-18092fff2622] Error from last host: compute (node
> compute): [u'Traceback (most recent call last):\n', u'  File
> "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 1787, in
> _do_build_and_run_instance\n    filter_properties)\n', u'  File
> "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 1985, in
> _build_and_run_instance\n    instance_uuid=instance.uuid,
> reason=six.text_type(e))\n', u'RescheduledException: Build of instance
> 0ba01247-5513-4c58-bf04-18092fff2622 was re-scheduled: Could
> not determine a suitable URL for the plugin\n']
> 2018-01-15 16:09:51.057 2567 WARNING nova.scheduler.utils
> [req-afff24dc-1ee0-469f-9d99-2abcb4810c7a
82858c289ca444bf90fcd41123d069ce
> 61b0b2b23b08419596bd923f2c544956 - - -] Failed to
> compute_task_build_instances: No valid host was found. There are not
enough
> hosts available.
> Traceback (most recent call last):
>
>   File "/usr/lib/python2.7/site-packages/oslo_messaging/rpc/server.py",
> line 199, in inner
>     return func(*args, **kwargs)
>
>   File "/usr/lib/python2.7/site-packages/nova/scheduler/manager.py", line
> 104, in select_destinations
>     dests = self.driver.select_destinations(ctxt, spec_obj)
>
>   File
> "/usr/lib/python2.7/site-packages/nova/scheduler/filter_scheduler.py",
line
> 74, in select_destinations
>     raise exception.NoValidHost(reason=reason)
>
> NoValidHost: No valid host was found. There are not enough hosts
available.
>
> 2018-01-15 16:09:51.057 2567 WARNING nova.scheduler.utils
> [req-afff24dc-1ee0-469f-9d99-2abcb4810c7a
82858c289ca444bf90fcd41123d069ce
> 61b0b2b23b08419596bd923f2c544956 - - -] [instance:
> 0ba01247-5513-4c58-bf04-18092fff2622] Setting instance to ERROR state.
>
> Any tips how to resolve this?
>
> Thanks



--
Eugen Block                             voice   : +49-40-559 51 75
<+49%2040%205595175>
NDE Netzdesign und -entwicklung AG      fax     : +49-40-559 51 77
<+49%2040%205595177>
Postfach 61 03 15
D-22423 Hamburg                         e-mail  : ebl...@nde.ag

         Vorsitzende des Aufsichtsrates: Angelika Mozdzen
           Sitz und Registergericht: Hamburg, HRB 90934
                   Vorstand: Jens-U. Mozdzen
                    USt-IdNr. DE 814 013 983


_______________________________________________
Mailing list:
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to     : openstack@lists.openstack.org
Unsubscribe :
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack




--
Eugen Block                             voice   : +49-40-559 51 75
NDE Netzdesign und -entwicklung AG      fax     : +49-40-559 51 77
Postfach 61 03 15
D-22423 Hamburg                         e-mail  : ebl...@nde.ag

        Vorsitzende des Aufsichtsrates: Angelika Mozdzen
          Sitz und Registergericht: Hamburg, HRB 90934
                  Vorstand: Jens-U. Mozdzen
                   USt-IdNr. DE 814 013 983


_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to     : openstack@lists.openstack.org
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack

Reply via email to