Which request generated this trace?
Is it  nova-compute log?

> On 10 Jul 2015, at 07:13, Mario Codeniera <[email protected]> wrote:
> 
> Hi,
> 
> It is my first time here. I am just having an issue regarding with my 
> configuration with the OpenStack which works perfectly for the cinder and the 
> glance based on Kilo release in CentOS 7. I am based my documentation on this 
> rbd-opeenstack manual.
> 
> 
> If I enable my rbd in the nova.conf it generates error like the following in 
> the dashboard as the logs don't have any errors:
> 
> Internal Server Error (HTTP 500) (Request-ID: 
> req-231347dd-f14c-4f97-8a1d-851a149b037c)
> Code
> 500
> Details
> File "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 343, in 
> decorated_function return function(self, context, *args, **kwargs) File 
> "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 2737, in 
> terminate_instance do_terminate_instance(instance, bdms) File 
> "/usr/lib/python2.7/site-packages/oslo_concurrency/lockutils.py", line 445, 
> in inner return f(*args, **kwargs) File 
> "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 2735, in 
> do_terminate_instance self._set_instance_error_state(context, instance) File 
> "/usr/lib/python2.7/site-packages/oslo_utils/excutils.py", line 85, in 
> __exit__ six.reraise(self.type_, self.value, self.tb) File 
> "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 2725, in 
> do_terminate_instance self._delete_instance(context, instance, bdms, quotas) 
> File "/usr/lib/python2.7/site-packages/nova/hooks.py", line 149, in inner rv 
> = f(*args, **kwargs) File 
> "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 2694, in 
> _delete_instance quotas.rollback() File 
> "/usr/lib/python2.7/site-packages/oslo_utils/excutils.py", line 85, in 
> __exit__ six.reraise(self.type_, self.value, self.tb) File 
> "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 2664, in 
> _delete_instance self._shutdown_instance(context, instance, bdms) File 
> "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 2604, in 
> _shutdown_instance self.volume_api.detach(context, bdm.volume_id) File 
> "/usr/lib/python2.7/site-packages/nova/volume/cinder.py", line 214, in 
> wrapper res = method(self, ctx, volume_id, *args, **kwargs) File 
> "/usr/lib/python2.7/site-packages/nova/volume/cinder.py", line 365, in detach 
> cinderclient(context).volumes.detach(volume_id) File 
> "/usr/lib/python2.7/site-packages/cinderclient/v2/volumes.py", line 334, in 
> detach return self._action('os-detach', volume) File 
> "/usr/lib/python2.7/site-packages/cinderclient/v2/volumes.py", line 311, in 
> _action return self.api.client.post(url, body=body) File 
> "/usr/lib/python2.7/site-packages/cinderclient/client.py", line 91, in post 
> return self._cs_request(url, 'POST', **kwargs) File 
> "/usr/lib/python2.7/site-packages/cinderclient/client.py", line 85, in 
> _cs_request return self.request(url, method, **kwargs) File 
> "/usr/lib/python2.7/site-packages/cinderclient/client.py", line 80, in 
> request return super(SessionClient, self).request(*args, **kwargs) File 
> "/usr/lib/python2.7/site-packages/keystoneclient/adapter.py", line 206, in 
> request resp = super(LegacyJsonAdapter, self).request(*args, **kwargs) File 
> "/usr/lib/python2.7/site-packages/keystoneclient/adapter.py", line 95, in 
> request return self.session.request(url, method, **kwargs) File 
> "/usr/lib/python2.7/site-packages/keystoneclient/utils.py", line 318, in 
> inner return func(*args, **kwargs) File 
> "/usr/lib/python2.7/site-packages/keystoneclient/session.py", line 397, in 
> request raise exceptions.from_response(resp, method, url)
> Created
> 10 Jul 2015, 4:40 a.m.
> 
> 
> Again if disable I able to work but it is generated on the compute node, as I 
> observe too it doesn't display the hypervisor of the compute nodes, or maybe 
> it is related.
> 
> It was working on Juno before, but there are unexpected rework as the network 
> infrastructure was change which the I rerun the script and found lots of 
> conflicts et al as I run before using qemu-img-rhev qemu-kvm-rhev from OVirt 
> but seems the new hammer (Ceph repository) solve the issue.
> 
> Hope someone can enlighten.
> 
> Thanks,
> Mario
> 
> 
> 
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Cheers.
––––
Sébastien Han
Senior Cloud Architect

"Always give 100%. Unless you're giving blood."

Mail: [email protected]
Address: 11 bis, rue Roquépine - 75008 Paris

Attachment: signature.asc
Description: Message signed with OpenPGP using GPGMail

_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to