Yes: nova-manage service list Binary Host Zone Status State Updated_At nova-cert hemera internal enabled :-) 2013-08-19 22:20:14 nova-scheduler hemera internal enabled :-) 2013-08-19 22:20:14 nova-consoleauth hemera internal enabled :-) 2013-08-19 22:20:17 nova-conductor hemera internal enabled :-) 2013-08-19 22:20:14 nova-compute caos nova enabled :-) 2013-08-19 22:20:17 nova-compute tiresias nova enabled :-) 2013-08-19 22:20:17
2013/8/19 Razique Mahroua <[email protected]> > Can you run nova-manage service list ? > > *Razique Mahroua** - **Nuage & Co* > [email protected] > Tel : +33 9 72 37 94 15 > > > Le 19 août 2013 à 22:27, Guilherme Russi <[email protected]> a > écrit : > > Hello again Razique, I was trying to migrate through the terminal and I'm > getting this line from my controller node: > > nova live-migration f18b6d2b-f685-4476-a7c0-e735db5113cf 192.168.3.3 > ERROR: Compute service of 192.168.3.3 is unavailable at this time. (HTTP > 400) (Request-ID: req-454960fb-6652-46ed-9080-6d23b384cbb3) > > Haven't found anything yet. Any ideas? > > Regards. > > Guilherme. > > > 2013/8/19 Guilherme Russi <[email protected]> > >> I'm migrating it through the dashboard with the admin user. Is it wrong? >> >> Thank you. >> >> Guilherme. >> >> >> 2013/8/19 Razique Mahroua <[email protected]> >> >>> quick question, how do you migrate the instance ? and does it belong to >>> the tenant? >>> >>> >>> Le 19 août 2013 à 15:36, Guilherme Russi <[email protected]> a >>> écrit : >>> >>> The VM looks to be there: >>> >>> du -h instances/ >>> 4,0K instances/keys >>> 1,9G instances/_base >>> 0 instances/locks >>> 310M instances/edca1460-9c97-475e-997b-3266b235e797 >>> 260M instances/0fe59b28-04a3-41bb-9926-b7d139a70548 >>> 323M instances/aeda1c56-72ba-40a7-9857-01a622540505 >>> 7,0M instances/f4af662d-ae55-4346-82cf-a997192706b5 >>> 264M instances/483f98e3-8ef5-43e2-8c3a-def55abdabcd >>> 268M instances/eeb81e32-32ce-49fa-8249-afc572f26233 >>> 268M instances/c0224f0d-d7e8-4621-8c1a-af68740a3267 >>> 188M instances/72ec37a3-b209-4729-b628-005fdcea5a3c >>> 12M instances/2c7c123b-456a-4c70-a991-80efee541324 >>> 11M instances/c9e1c5ed-a108-4196-bfbc-24495e2e71bd >>> 6,5M instances/96d8f2de-a72d-40b5-b083-1f5e511f91ce >>> 531M instances/1b35a4a6-e158-48d3-a804-44b05ec171d0 >>> 529M instances/1a2b6146-6ef0-4d76-a603-e4715b4dfb62 >>> 514M instances/f6cb9f81-9281-419f-9e20-adaa6a07cea3 >>> 2,7G instances/75e479cf-37a4-4511-b2a5-d4f3027702d4_resize >>> 2,4G instances/75e479cf-37a4-4511-b2a5-d4f3027702d4 >>> 11G instances/ >>> >>> >>> It's still running, no Error messagem from dashboard (it is showing >>> Status: Resize/Migrate - Task: Resizing or Migrating - Power State: >>> Running). The logs are not showing that permission error anymore. >>> Should I wait to see if it will migrate or returns another error status? >>> One thing is true, the machine is really slow now :P >>> >>> Regards. >>> >>> Guilherme. >>> >>> >>> 2013/8/19 Razique Mahroua <[email protected]> >>> >>>> Looks like the file doesn't exist. >>>> Can you spawn a new instance and try to migrate it again? >>>> thanks >>>> >>>> *Razique Mahroua** - **Nuage & Co* >>>> [email protected] >>>> Tel : +33 9 72 37 94 15 >>>> >>>> <NUAGECO-LOGO-Fblan_petit.jpg> >>>> >>>> Le 19 août 2013 à 13:41, Guilherme Russi <[email protected]> >>>> a écrit : >>>> >>>> Hello Peter, >>>> >>>> Thank you for your help too :) >>>> I've changed the permission and now I'm getting another error when I >>>> try to migrate: >>>> >>>> cat /var/log/nova/nova-compute.log | grep -E ERROR >>>> 2013-08-19 08:17:19.990 12667 ERROR nova.manager [-] Error during >>>> ComputeManager._run_image_cache_manager_pass: Unexpected error while >>>> running command. >>>> 2013-08-19 08:23:40.309 12667 ERROR nova.network.quantumv2 [-] >>>> _get_auth_token() failed >>>> 2013-08-19 08:24:27.013 12667 ERROR nova.virt.libvirt.driver [-] >>>> Getting disk size of instance-000000bc: [Errno 2] No such file or >>>> directory: >>>> '/var/lib/nova/instances/75e479cf-37a4-4511-b2a5-d4f3027702d4/disk' >>>> >>>> >>>> cat /var/log/nova/nova-compute.log | grep -E TRACE >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Traceback (most >>>> recent call last): >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>>> "/usr/lib/python2.7/dist-packages/nova/manager.py", line 241, in >>>> periodic_tasks >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager task(self, >>>> context) >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>>> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 4020, in >>>> _run_image_cache_manager_pass >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager >>>> self.driver.manage_image_cache(context, filtered_instances) >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py", line 3401, >>>> in manage_image_cache >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager >>>> self.image_cache_manager.verify_base_images(context, all_instances) >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/imagecache.py", line >>>> 588, in verify_base_images >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager >>>> self._handle_base_image(img, base_file) >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/imagecache.py", line >>>> 553, in _handle_base_image >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager >>>> virtutils.chown(base_file, os.getuid()) >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/utils.py", line 449, in >>>> chown >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager execute('chown', >>>> owner, path, run_as_root=True) >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/utils.py", line 47, in >>>> execute >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager return >>>> utils.execute(*args, **kwargs) >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>>> "/usr/lib/python2.7/dist-packages/nova/utils.py", line 239, in execute >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager cmd=' '.join(cmd)) >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager >>>> ProcessExecutionError: Unexpected error while running command. >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Command: sudo >>>> nova-rootwrap /etc/nova/rootwrap.conf chown 123 >>>> /var/lib/nova/instances/_base/b0d90eb60f048f6de16c3efbc0eb4c16cbac77a6 >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Exit code: 1 >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Stdout: '' >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Stderr: 'Traceback >>>> (most recent call last):\n File "/usr/bin/nova-rootwrap", line 108, in >>>> <module>\n os.getlogin(), pwd.getpwuid(os.getuid())[0],\nOSError: [Errno >>>> 22] Invalid argument\n' >>>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager >>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 Traceback >>>> (most recent call last): >>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 File >>>> "/usr/lib/python2.7/dist-packages/nova/network/quantumv2/__init__.py", line >>>> 40, in _get_auth_token >>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 >>>> httpclient.authenticate() >>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 File >>>> "/usr/lib/python2.7/dist-packages/quantumclient/client.py", line 198, in >>>> authenticate >>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 raise >>>> exceptions.Unauthorized(message=body) >>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 >>>> Unauthorized: Request Timeout >>>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 >>>> >>>> Any ideas? >>>> >>>> Regards. >>>> >>>> Guilherme. >>>> >>>> >>>> 2013/8/19 Peter Funk <[email protected]> >>>> >>>>> Guilherme Russi wrote 16.08.2013 um 09:53: >>>>> > Here is my directory configs: >>>>> > >>>>> > drwxrwxrwx 2 nova nova 76 Ago 15 14:54 .ssh >>>>> > >>>>> > I made chmod a+xwr at the folder at all computers, I made it because >>>>> I >>>>> > couldn't find the reason I was no allowed to migrate, can it be like >>>>> that? >>>>> >>>>> For SSH to work the permissions of $HOME/.ssh should look like this: >>>>> drwx------ 2 nova nova 76 Ago 15 14:54 .ssh >>>>> >>>>> Use: >>>>> chmod -R go-rwx .ssh >>>>> >>>>> Regards, Peter Funk >>>>> -- >>>>> Peter Funk, home: ✉Oldenburger Str.86, D-27777 Ganderkesee >>>>> mobile:+49-179-640-8878 phone:+49-421-20419-0 < >>>>> http://www.artcom-gmbh.de/> >>>>> office: ArtCom GmbH, ✉Haferwende 2, D-28357 Bremen, Germany >>>>> >>>>> _______________________________________________ >>>>> Mailing list: >>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>>>> Post to : [email protected] >>>>> Unsubscribe : >>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>>>> >>>> >>>> _______________________________________________ >>>> Mailing list: >>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>>> Post to : [email protected] >>>> Unsubscribe : >>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>>> >>>> >>>> >>> >>> >> > >
<<NUAGECO-LOGO-Fblan_petit.jpg>>
_______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : [email protected] Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
