Re: Can't destroy system VMs

2013-11-07 Thread Daan Hoogland
Sorry Carlos,

family and $dayjob kept me from looking at this anymore. Glad you got
it working.

Daan

On Wed, Nov 6, 2013 at 8:08 PM, Carlos Reategui car...@reategui.com wrote:
 Since I didn't hear back I bit the bullet and decided to make the DB
 changes and I am back up and running.  Woohoo.  I'll post the specifics in
 the other thread I started about the mshost id.


 On Tue, Nov 5, 2013 at 11:36 AM, Carlos Reategui car...@reategui.comwrote:

 Hi Daan,

 The mshost thing I mentioned previously has been bugging me and as I look
 through the db I think I have a problem here.  I am guessing I have a
 botched 4.1.0 to 4.1.1 upgrade that I did not notice when I shut things
 down.

 The mshost table is showing 2 management servers that are supposedly on
 the same machine listening on the same port which is not really possible.
  So really only one of the 2 is up (I believe that would be the 4.1.1
 version). The logs do seem to indicate something along those lines where
 the 2 management servers are failing to communicate with each other.

 If I look at the host table I see my 2 hosts but they say they are on
 version 4.1.0 and are pointing to management server id of the 4.1.0 line in
 the mshost table.

 I am wondering if this is the problem and why their is no communication to
 my hosts from the management server.

 Anyone know the schema well enough to help me update some stuff and
 recover my installation?  Would an upgrade to 4.2 possibly cleanup my DB?

 thanks,
 Carlos




 On Tue, Nov 5, 2013 at 10:59 AM, Carlos Reategui car...@reategui.comwrote:

 Hi Daan,

 I have been looking through my logs and can't figure out why the ssvm
 does not want to start.

 Here is a link to my latest logs:
 http://reategui.com/cloudstack/management-server-131105.log

 Maybe someone else can spot something.

 There is nothing on either the XenServer hosts logs.

 thanks,
 Carlos


 On Tue, Nov 5, 2013 at 12:55 AM, Daan Hoogland 
 daan.hoogl...@gmail.comwrote:

 Carlos,

 It makes no sense trying to start the vms if not the ssvm is started
 first.

 I would say stop the system, clear the logs and start again. Then
 analyse the logs to see what is wrong with the secondary storage vm.

 On Tue, Nov 5, 2013 at 2:08 AM, Carlos Reategui car...@reategui.com
 wrote:
  Hi Daan,
  There is no trace of CS communicating with either of my XS hosts in the
  SMLog file or the xensource.log files when I try to start one of my
 stopped
  instances.
 
  thanks,
  Carlos
 
 
 
  On Sat, Nov 2, 2013 at 4:50 AM, Daan Hoogland daan.hoogl...@gmail.com
 wrote:
 
  H Carlos,
 
  I wouldn't know about the mshost and mshost_peer but they seem
  alright. The expunging ssvm are a real problem. I think Ahmed
  mentioned that as well. You are having problems getting your ssvm in
  the air. All other things are consequential. This is the first thing
  to look at. Check the logs and the SMLog on the hypervisor to see what
  happens when cs tries to start a ssvm. By your db you can see that it
  tried to start it several times. Those are the expunging vm_instance
  entries.
 
  regards,
  Daan
 
  On Fri, Nov 1, 2013 at 9:10 PM, Carlos Reategui car...@reategui.com
  wrote:
   Hi Daan  :)
  
   Primary and Secondary storage are on the MS.  Primary is on a
 different
   subnet and it is currently mounted (and visible as an NFS SR from
   XenCenter) and working fine from the xen machines.  I have also
 been able
   to mount Secondary without a problem.
  
   Here are my current system VMs:
   mysql select
   id,name,instance_name,state,host_id,type,vm_type,desired_state from
   vm_instance where type!='User';
  
 
 ++-+---+---+-+++---+
   | id | name| instance_name | state | host_id | type
|
   vm_type| desired_state |
  
 
 ++-+---+---+-+++---+
   |  2 | v-2-VM  | v-2-VM| Stopped   |   1 | ConsoleProxy
|
   ConsoleProxy   | NULL  |
   |  5 | r-5-VM  | r-5-VM| Stopped   |   2 | DomainRouter
|
   DomainRouter   | NULL  |
   |  1 | s-1-VM  | s-1-VM| Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   |  3 | s-3-VM  | s-3-VM| Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 52 | s-52-VM | s-52-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 53 | s-53-VM | s-53-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 54 | s-54-VM | s-54-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 55 | s-55-VM | s-55-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 56 | s-56-VM | s-56-VM   | Expunging |NULL |
  

Re: Can't destroy system VMs

2013-11-06 Thread Carlos Reategui
Since I didn't hear back I bit the bullet and decided to make the DB
changes and I am back up and running.  Woohoo.  I'll post the specifics in
the other thread I started about the mshost id.


On Tue, Nov 5, 2013 at 11:36 AM, Carlos Reategui car...@reategui.comwrote:

 Hi Daan,

 The mshost thing I mentioned previously has been bugging me and as I look
 through the db I think I have a problem here.  I am guessing I have a
 botched 4.1.0 to 4.1.1 upgrade that I did not notice when I shut things
 down.

 The mshost table is showing 2 management servers that are supposedly on
 the same machine listening on the same port which is not really possible.
  So really only one of the 2 is up (I believe that would be the 4.1.1
 version). The logs do seem to indicate something along those lines where
 the 2 management servers are failing to communicate with each other.

 If I look at the host table I see my 2 hosts but they say they are on
 version 4.1.0 and are pointing to management server id of the 4.1.0 line in
 the mshost table.

 I am wondering if this is the problem and why their is no communication to
 my hosts from the management server.

 Anyone know the schema well enough to help me update some stuff and
 recover my installation?  Would an upgrade to 4.2 possibly cleanup my DB?

 thanks,
 Carlos




 On Tue, Nov 5, 2013 at 10:59 AM, Carlos Reategui car...@reategui.comwrote:

 Hi Daan,

 I have been looking through my logs and can't figure out why the ssvm
 does not want to start.

 Here is a link to my latest logs:
 http://reategui.com/cloudstack/management-server-131105.log

 Maybe someone else can spot something.

 There is nothing on either the XenServer hosts logs.

 thanks,
 Carlos


 On Tue, Nov 5, 2013 at 12:55 AM, Daan Hoogland 
 daan.hoogl...@gmail.comwrote:

 Carlos,

 It makes no sense trying to start the vms if not the ssvm is started
 first.

 I would say stop the system, clear the logs and start again. Then
 analyse the logs to see what is wrong with the secondary storage vm.

 On Tue, Nov 5, 2013 at 2:08 AM, Carlos Reategui car...@reategui.com
 wrote:
  Hi Daan,
  There is no trace of CS communicating with either of my XS hosts in the
  SMLog file or the xensource.log files when I try to start one of my
 stopped
  instances.
 
  thanks,
  Carlos
 
 
 
  On Sat, Nov 2, 2013 at 4:50 AM, Daan Hoogland daan.hoogl...@gmail.com
 wrote:
 
  H Carlos,
 
  I wouldn't know about the mshost and mshost_peer but they seem
  alright. The expunging ssvm are a real problem. I think Ahmed
  mentioned that as well. You are having problems getting your ssvm in
  the air. All other things are consequential. This is the first thing
  to look at. Check the logs and the SMLog on the hypervisor to see what
  happens when cs tries to start a ssvm. By your db you can see that it
  tried to start it several times. Those are the expunging vm_instance
  entries.
 
  regards,
  Daan
 
  On Fri, Nov 1, 2013 at 9:10 PM, Carlos Reategui car...@reategui.com
  wrote:
   Hi Daan  :)
  
   Primary and Secondary storage are on the MS.  Primary is on a
 different
   subnet and it is currently mounted (and visible as an NFS SR from
   XenCenter) and working fine from the xen machines.  I have also
 been able
   to mount Secondary without a problem.
  
   Here are my current system VMs:
   mysql select
   id,name,instance_name,state,host_id,type,vm_type,desired_state from
   vm_instance where type!='User';
  
 
 ++-+---+---+-+++---+
   | id | name| instance_name | state | host_id | type
|
   vm_type| desired_state |
  
 
 ++-+---+---+-+++---+
   |  2 | v-2-VM  | v-2-VM| Stopped   |   1 | ConsoleProxy
|
   ConsoleProxy   | NULL  |
   |  5 | r-5-VM  | r-5-VM| Stopped   |   2 | DomainRouter
|
   DomainRouter   | NULL  |
   |  1 | s-1-VM  | s-1-VM| Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   |  3 | s-3-VM  | s-3-VM| Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 52 | s-52-VM | s-52-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 53 | s-53-VM | s-53-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 54 | s-54-VM | s-54-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 55 | s-55-VM | s-55-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 56 | s-56-VM | s-56-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
  
 
 ++-+---+---+-+++---+
  
  
  

Re: Can't destroy system VMs

2013-11-05 Thread Daan Hoogland
Carlos,

It makes no sense trying to start the vms if not the ssvm is started first.

I would say stop the system, clear the logs and start again. Then
analyse the logs to see what is wrong with the secondary storage vm.

On Tue, Nov 5, 2013 at 2:08 AM, Carlos Reategui car...@reategui.com wrote:
 Hi Daan,
 There is no trace of CS communicating with either of my XS hosts in the
 SMLog file or the xensource.log files when I try to start one of my stopped
 instances.

 thanks,
 Carlos



 On Sat, Nov 2, 2013 at 4:50 AM, Daan Hoogland daan.hoogl...@gmail.comwrote:

 H Carlos,

 I wouldn't know about the mshost and mshost_peer but they seem
 alright. The expunging ssvm are a real problem. I think Ahmed
 mentioned that as well. You are having problems getting your ssvm in
 the air. All other things are consequential. This is the first thing
 to look at. Check the logs and the SMLog on the hypervisor to see what
 happens when cs tries to start a ssvm. By your db you can see that it
 tried to start it several times. Those are the expunging vm_instance
 entries.

 regards,
 Daan

 On Fri, Nov 1, 2013 at 9:10 PM, Carlos Reategui car...@reategui.com
 wrote:
  Hi Daan  :)
 
  Primary and Secondary storage are on the MS.  Primary is on a different
  subnet and it is currently mounted (and visible as an NFS SR from
  XenCenter) and working fine from the xen machines.  I have also been able
  to mount Secondary without a problem.
 
  Here are my current system VMs:
  mysql select
  id,name,instance_name,state,host_id,type,vm_type,desired_state from
  vm_instance where type!='User';
 
 ++-+---+---+-+++---+
  | id | name| instance_name | state | host_id | type
   |
  vm_type| desired_state |
 
 ++-+---+---+-+++---+
  |  2 | v-2-VM  | v-2-VM| Stopped   |   1 | ConsoleProxy
   |
  ConsoleProxy   | NULL  |
  |  5 | r-5-VM  | r-5-VM| Stopped   |   2 | DomainRouter
   |
  DomainRouter   | NULL  |
  |  1 | s-1-VM  | s-1-VM| Expunging |NULL |
 SecondaryStorageVm |
  SecondaryStorageVm | NULL  |
  |  3 | s-3-VM  | s-3-VM| Expunging |NULL |
 SecondaryStorageVm |
  SecondaryStorageVm | NULL  |
  | 52 | s-52-VM | s-52-VM   | Expunging |NULL |
 SecondaryStorageVm |
  SecondaryStorageVm | NULL  |
  | 53 | s-53-VM | s-53-VM   | Expunging |NULL |
 SecondaryStorageVm |
  SecondaryStorageVm | NULL  |
  | 54 | s-54-VM | s-54-VM   | Expunging |NULL |
 SecondaryStorageVm |
  SecondaryStorageVm | NULL  |
  | 55 | s-55-VM | s-55-VM   | Expunging |NULL |
 SecondaryStorageVm |
  SecondaryStorageVm | NULL  |
  | 56 | s-56-VM | s-56-VM   | Expunging |NULL |
 SecondaryStorageVm |
  SecondaryStorageVm | NULL  |
 
 ++-+---+---+-+++---+
 
 
  I have been looking through the DB to see if anything sticks out.  I
 notice
  that the mshost and mshost_peer table have 2 entries but they are the
 same
  server.  Is that ok?  Looks like one is from the install of 4.1.0 an the
  other from after I upgraded to 4.1.1:
 
  mysql select * from mshost;
 
 ++-+---++---+-+-+--+-+-+-+
  | id | msid| runid | name   | state | version |
  service_ip  | service_port | last_update | removed | alert_count
 |
 
 ++-+---++---+-+-+--+-+-+-+
  |  1 | 233845174730255 | 1375383759777 | isbustor01 | Up| 4.1.0   |
  172.30.45.2 | 9090 | 2013-10-24 21:13:10 | NULL|   0
 |
  |  2 | 233845174730253 | 1383331508626 | isbustor01 | Up| 4.1.1   |
  172.30.45.2 | 9090 | 2013-11-01 20:00:04 | NULL|   0
 |
 
 ++-+---++---+-+-+--+-+-+-+
  2 rows in set (0.00 sec)
 
  mysql select * from mshost_peer;
 
 ++--+-+---++-+
  | id | owner_mshost | peer_mshost | peer_runid| peer_state |
  last_update |
 
 ++--+-+---++-+
  | 17 |1 |   1 | 1375383759777 | Up |
 2013-08-01
  19:02:49 |
  | 34 |2 |   2 | 1383331508626 | Up |
 2013-11-01
  18:45:19 |
 
 ++--+-+---++-+
 
 
  thanks,
  Carlos
 
 
 
  On Fri, Nov 1, 2013 at 12:17 PM, Daan 

Re: Can't destroy system VMs

2013-11-05 Thread Carlos Reategui
Hi Daan,

I have been looking through my logs and can't figure out why the ssvm does
not want to start.

Here is a link to my latest logs:
http://reategui.com/cloudstack/management-server-131105.log

Maybe someone else can spot something.

There is nothing on either the XenServer hosts logs.

thanks,
Carlos


On Tue, Nov 5, 2013 at 12:55 AM, Daan Hoogland daan.hoogl...@gmail.comwrote:

 Carlos,

 It makes no sense trying to start the vms if not the ssvm is started first.

 I would say stop the system, clear the logs and start again. Then
 analyse the logs to see what is wrong with the secondary storage vm.

 On Tue, Nov 5, 2013 at 2:08 AM, Carlos Reategui car...@reategui.com
 wrote:
  Hi Daan,
  There is no trace of CS communicating with either of my XS hosts in the
  SMLog file or the xensource.log files when I try to start one of my
 stopped
  instances.
 
  thanks,
  Carlos
 
 
 
  On Sat, Nov 2, 2013 at 4:50 AM, Daan Hoogland daan.hoogl...@gmail.com
 wrote:
 
  H Carlos,
 
  I wouldn't know about the mshost and mshost_peer but they seem
  alright. The expunging ssvm are a real problem. I think Ahmed
  mentioned that as well. You are having problems getting your ssvm in
  the air. All other things are consequential. This is the first thing
  to look at. Check the logs and the SMLog on the hypervisor to see what
  happens when cs tries to start a ssvm. By your db you can see that it
  tried to start it several times. Those are the expunging vm_instance
  entries.
 
  regards,
  Daan
 
  On Fri, Nov 1, 2013 at 9:10 PM, Carlos Reategui car...@reategui.com
  wrote:
   Hi Daan  :)
  
   Primary and Secondary storage are on the MS.  Primary is on a
 different
   subnet and it is currently mounted (and visible as an NFS SR from
   XenCenter) and working fine from the xen machines.  I have also been
 able
   to mount Secondary without a problem.
  
   Here are my current system VMs:
   mysql select
   id,name,instance_name,state,host_id,type,vm_type,desired_state from
   vm_instance where type!='User';
  
 
 ++-+---+---+-+++---+
   | id | name| instance_name | state | host_id | type
|
   vm_type| desired_state |
  
 
 ++-+---+---+-+++---+
   |  2 | v-2-VM  | v-2-VM| Stopped   |   1 | ConsoleProxy
|
   ConsoleProxy   | NULL  |
   |  5 | r-5-VM  | r-5-VM| Stopped   |   2 | DomainRouter
|
   DomainRouter   | NULL  |
   |  1 | s-1-VM  | s-1-VM| Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   |  3 | s-3-VM  | s-3-VM| Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 52 | s-52-VM | s-52-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 53 | s-53-VM | s-53-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 54 | s-54-VM | s-54-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 55 | s-55-VM | s-55-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 56 | s-56-VM | s-56-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
  
 
 ++-+---+---+-+++---+
  
  
   I have been looking through the DB to see if anything sticks out.  I
  notice
   that the mshost and mshost_peer table have 2 entries but they are the
  same
   server.  Is that ok?  Looks like one is from the install of 4.1.0 an
 the
   other from after I upgraded to 4.1.1:
  
   mysql select * from mshost;
  
 
 ++-+---++---+-+-+--+-+-+-+
   | id | msid| runid | name   | state | version
 |
   service_ip  | service_port | last_update | removed |
 alert_count
  |
  
 
 ++-+---++---+-+-+--+-+-+-+
   |  1 | 233845174730255 | 1375383759777 | isbustor01 | Up| 4.1.0
 |
   172.30.45.2 | 9090 | 2013-10-24 21:13:10 | NULL|
   0
  |
   |  2 | 233845174730253 | 1383331508626 | isbustor01 | Up| 4.1.1
 |
   172.30.45.2 | 9090 | 2013-11-01 20:00:04 | NULL|
   0
  |
  
 
 ++-+---++---+-+-+--+-+-+-+
   2 rows in set (0.00 sec)
  
   mysql select * from mshost_peer;
  
 
 ++--+-+---++-+
   | id | owner_mshost | 

Re: Can't destroy system VMs

2013-11-05 Thread Carlos Reategui
Hi Daan,

The mshost thing I mentioned previously has been bugging me and as I look
through the db I think I have a problem here.  I am guessing I have a
botched 4.1.0 to 4.1.1 upgrade that I did not notice when I shut things
down.

The mshost table is showing 2 management servers that are supposedly on the
same machine listening on the same port which is not really possible.  So
really only one of the 2 is up (I believe that would be the 4.1.1 version).
The logs do seem to indicate something along those lines where the 2
management servers are failing to communicate with each other.

If I look at the host table I see my 2 hosts but they say they are on
version 4.1.0 and are pointing to management server id of the 4.1.0 line in
the mshost table.

I am wondering if this is the problem and why their is no communication to
my hosts from the management server.

Anyone know the schema well enough to help me update some stuff and recover
my installation?  Would an upgrade to 4.2 possibly cleanup my DB?

thanks,
Carlos




On Tue, Nov 5, 2013 at 10:59 AM, Carlos Reategui car...@reategui.comwrote:

 Hi Daan,

 I have been looking through my logs and can't figure out why the ssvm does
 not want to start.

 Here is a link to my latest logs:
 http://reategui.com/cloudstack/management-server-131105.log

 Maybe someone else can spot something.

 There is nothing on either the XenServer hosts logs.

 thanks,
 Carlos


 On Tue, Nov 5, 2013 at 12:55 AM, Daan Hoogland daan.hoogl...@gmail.comwrote:

 Carlos,

 It makes no sense trying to start the vms if not the ssvm is started
 first.

 I would say stop the system, clear the logs and start again. Then
 analyse the logs to see what is wrong with the secondary storage vm.

 On Tue, Nov 5, 2013 at 2:08 AM, Carlos Reategui car...@reategui.com
 wrote:
  Hi Daan,
  There is no trace of CS communicating with either of my XS hosts in the
  SMLog file or the xensource.log files when I try to start one of my
 stopped
  instances.
 
  thanks,
  Carlos
 
 
 
  On Sat, Nov 2, 2013 at 4:50 AM, Daan Hoogland daan.hoogl...@gmail.com
 wrote:
 
  H Carlos,
 
  I wouldn't know about the mshost and mshost_peer but they seem
  alright. The expunging ssvm are a real problem. I think Ahmed
  mentioned that as well. You are having problems getting your ssvm in
  the air. All other things are consequential. This is the first thing
  to look at. Check the logs and the SMLog on the hypervisor to see what
  happens when cs tries to start a ssvm. By your db you can see that it
  tried to start it several times. Those are the expunging vm_instance
  entries.
 
  regards,
  Daan
 
  On Fri, Nov 1, 2013 at 9:10 PM, Carlos Reategui car...@reategui.com
  wrote:
   Hi Daan  :)
  
   Primary and Secondary storage are on the MS.  Primary is on a
 different
   subnet and it is currently mounted (and visible as an NFS SR from
   XenCenter) and working fine from the xen machines.  I have also been
 able
   to mount Secondary without a problem.
  
   Here are my current system VMs:
   mysql select
   id,name,instance_name,state,host_id,type,vm_type,desired_state from
   vm_instance where type!='User';
  
 
 ++-+---+---+-+++---+
   | id | name| instance_name | state | host_id | type
|
   vm_type| desired_state |
  
 
 ++-+---+---+-+++---+
   |  2 | v-2-VM  | v-2-VM| Stopped   |   1 | ConsoleProxy
|
   ConsoleProxy   | NULL  |
   |  5 | r-5-VM  | r-5-VM| Stopped   |   2 | DomainRouter
|
   DomainRouter   | NULL  |
   |  1 | s-1-VM  | s-1-VM| Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   |  3 | s-3-VM  | s-3-VM| Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 52 | s-52-VM | s-52-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 53 | s-53-VM | s-53-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 54 | s-54-VM | s-54-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 55 | s-55-VM | s-55-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
   | 56 | s-56-VM | s-56-VM   | Expunging |NULL |
  SecondaryStorageVm |
   SecondaryStorageVm | NULL  |
  
 
 ++-+---+---+-+++---+
  
  
   I have been looking through the DB to see if anything sticks out.  I
  notice
   that the mshost and mshost_peer table have 2 entries but they are the
  same
   server.  Is that ok?  Looks like one is from the install of 4.1.0 an
 the
   other from after I upgraded to 4.1.1:
  
   mysql 

Re: Can't destroy system VMs

2013-11-04 Thread Carlos Reategui
Hi Daan,
There is no trace of CS communicating with either of my XS hosts in the
SMLog file or the xensource.log files when I try to start one of my stopped
instances.

thanks,
Carlos



On Sat, Nov 2, 2013 at 4:50 AM, Daan Hoogland daan.hoogl...@gmail.comwrote:

 H Carlos,

 I wouldn't know about the mshost and mshost_peer but they seem
 alright. The expunging ssvm are a real problem. I think Ahmed
 mentioned that as well. You are having problems getting your ssvm in
 the air. All other things are consequential. This is the first thing
 to look at. Check the logs and the SMLog on the hypervisor to see what
 happens when cs tries to start a ssvm. By your db you can see that it
 tried to start it several times. Those are the expunging vm_instance
 entries.

 regards,
 Daan

 On Fri, Nov 1, 2013 at 9:10 PM, Carlos Reategui car...@reategui.com
 wrote:
  Hi Daan  :)
 
  Primary and Secondary storage are on the MS.  Primary is on a different
  subnet and it is currently mounted (and visible as an NFS SR from
  XenCenter) and working fine from the xen machines.  I have also been able
  to mount Secondary without a problem.
 
  Here are my current system VMs:
  mysql select
  id,name,instance_name,state,host_id,type,vm_type,desired_state from
  vm_instance where type!='User';
 
 ++-+---+---+-+++---+
  | id | name| instance_name | state | host_id | type
   |
  vm_type| desired_state |
 
 ++-+---+---+-+++---+
  |  2 | v-2-VM  | v-2-VM| Stopped   |   1 | ConsoleProxy
   |
  ConsoleProxy   | NULL  |
  |  5 | r-5-VM  | r-5-VM| Stopped   |   2 | DomainRouter
   |
  DomainRouter   | NULL  |
  |  1 | s-1-VM  | s-1-VM| Expunging |NULL |
 SecondaryStorageVm |
  SecondaryStorageVm | NULL  |
  |  3 | s-3-VM  | s-3-VM| Expunging |NULL |
 SecondaryStorageVm |
  SecondaryStorageVm | NULL  |
  | 52 | s-52-VM | s-52-VM   | Expunging |NULL |
 SecondaryStorageVm |
  SecondaryStorageVm | NULL  |
  | 53 | s-53-VM | s-53-VM   | Expunging |NULL |
 SecondaryStorageVm |
  SecondaryStorageVm | NULL  |
  | 54 | s-54-VM | s-54-VM   | Expunging |NULL |
 SecondaryStorageVm |
  SecondaryStorageVm | NULL  |
  | 55 | s-55-VM | s-55-VM   | Expunging |NULL |
 SecondaryStorageVm |
  SecondaryStorageVm | NULL  |
  | 56 | s-56-VM | s-56-VM   | Expunging |NULL |
 SecondaryStorageVm |
  SecondaryStorageVm | NULL  |
 
 ++-+---+---+-+++---+
 
 
  I have been looking through the DB to see if anything sticks out.  I
 notice
  that the mshost and mshost_peer table have 2 entries but they are the
 same
  server.  Is that ok?  Looks like one is from the install of 4.1.0 an the
  other from after I upgraded to 4.1.1:
 
  mysql select * from mshost;
 
 ++-+---++---+-+-+--+-+-+-+
  | id | msid| runid | name   | state | version |
  service_ip  | service_port | last_update | removed | alert_count
 |
 
 ++-+---++---+-+-+--+-+-+-+
  |  1 | 233845174730255 | 1375383759777 | isbustor01 | Up| 4.1.0   |
  172.30.45.2 | 9090 | 2013-10-24 21:13:10 | NULL|   0
 |
  |  2 | 233845174730253 | 1383331508626 | isbustor01 | Up| 4.1.1   |
  172.30.45.2 | 9090 | 2013-11-01 20:00:04 | NULL|   0
 |
 
 ++-+---++---+-+-+--+-+-+-+
  2 rows in set (0.00 sec)
 
  mysql select * from mshost_peer;
 
 ++--+-+---++-+
  | id | owner_mshost | peer_mshost | peer_runid| peer_state |
  last_update |
 
 ++--+-+---++-+
  | 17 |1 |   1 | 1375383759777 | Up |
 2013-08-01
  19:02:49 |
  | 34 |2 |   2 | 1383331508626 | Up |
 2013-11-01
  18:45:19 |
 
 ++--+-+---++-+
 
 
  thanks,
  Carlos
 
 
 
  On Fri, Nov 1, 2013 at 12:17 PM, Daan Hoogland daan.hoogl...@gmail.com
 wrote:
 
  h Carlos,
 
  the states of virtual machines, especially the entries for the ssvm.
  if it is not Stopped. set it to stopped and check if your storage,
  both primary and secondary, are reachable botjh from the ms and the
  hosts.
 
  Daan (double a!)
 
  On Fri, 

Re: Can't destroy system VMs

2013-11-02 Thread Daan Hoogland
H Carlos,

I wouldn't know about the mshost and mshost_peer but they seem
alright. The expunging ssvm are a real problem. I think Ahmed
mentioned that as well. You are having problems getting your ssvm in
the air. All other things are consequential. This is the first thing
to look at. Check the logs and the SMLog on the hypervisor to see what
happens when cs tries to start a ssvm. By your db you can see that it
tried to start it several times. Those are the expunging vm_instance
entries.

regards,
Daan

On Fri, Nov 1, 2013 at 9:10 PM, Carlos Reategui car...@reategui.com wrote:
 Hi Daan  :)

 Primary and Secondary storage are on the MS.  Primary is on a different
 subnet and it is currently mounted (and visible as an NFS SR from
 XenCenter) and working fine from the xen machines.  I have also been able
 to mount Secondary without a problem.

 Here are my current system VMs:
 mysql select
 id,name,instance_name,state,host_id,type,vm_type,desired_state from
 vm_instance where type!='User';
 ++-+---+---+-+++---+
 | id | name| instance_name | state | host_id | type   |
 vm_type| desired_state |
 ++-+---+---+-+++---+
 |  2 | v-2-VM  | v-2-VM| Stopped   |   1 | ConsoleProxy   |
 ConsoleProxy   | NULL  |
 |  5 | r-5-VM  | r-5-VM| Stopped   |   2 | DomainRouter   |
 DomainRouter   | NULL  |
 |  1 | s-1-VM  | s-1-VM| Expunging |NULL | SecondaryStorageVm |
 SecondaryStorageVm | NULL  |
 |  3 | s-3-VM  | s-3-VM| Expunging |NULL | SecondaryStorageVm |
 SecondaryStorageVm | NULL  |
 | 52 | s-52-VM | s-52-VM   | Expunging |NULL | SecondaryStorageVm |
 SecondaryStorageVm | NULL  |
 | 53 | s-53-VM | s-53-VM   | Expunging |NULL | SecondaryStorageVm |
 SecondaryStorageVm | NULL  |
 | 54 | s-54-VM | s-54-VM   | Expunging |NULL | SecondaryStorageVm |
 SecondaryStorageVm | NULL  |
 | 55 | s-55-VM | s-55-VM   | Expunging |NULL | SecondaryStorageVm |
 SecondaryStorageVm | NULL  |
 | 56 | s-56-VM | s-56-VM   | Expunging |NULL | SecondaryStorageVm |
 SecondaryStorageVm | NULL  |
 ++-+---+---+-+++---+


 I have been looking through the DB to see if anything sticks out.  I notice
 that the mshost and mshost_peer table have 2 entries but they are the same
 server.  Is that ok?  Looks like one is from the install of 4.1.0 an the
 other from after I upgraded to 4.1.1:

 mysql select * from mshost;
 ++-+---++---+-+-+--+-+-+-+
 | id | msid| runid | name   | state | version |
 service_ip  | service_port | last_update | removed | alert_count |
 ++-+---++---+-+-+--+-+-+-+
 |  1 | 233845174730255 | 1375383759777 | isbustor01 | Up| 4.1.0   |
 172.30.45.2 | 9090 | 2013-10-24 21:13:10 | NULL|   0 |
 |  2 | 233845174730253 | 1383331508626 | isbustor01 | Up| 4.1.1   |
 172.30.45.2 | 9090 | 2013-11-01 20:00:04 | NULL|   0 |
 ++-+---++---+-+-+--+-+-+-+
 2 rows in set (0.00 sec)

 mysql select * from mshost_peer;
 ++--+-+---++-+
 | id | owner_mshost | peer_mshost | peer_runid| peer_state |
 last_update |
 ++--+-+---++-+
 | 17 |1 |   1 | 1375383759777 | Up | 2013-08-01
 19:02:49 |
 | 34 |2 |   2 | 1383331508626 | Up | 2013-11-01
 18:45:19 |
 ++--+-+---++-+


 thanks,
 Carlos



 On Fri, Nov 1, 2013 at 12:17 PM, Daan Hoogland daan.hoogl...@gmail.comwrote:

 h Carlos,

 the states of virtual machines, especially the entries for the ssvm.
 if it is not Stopped. set it to stopped and check if your storage,
 both primary and secondary, are reachable botjh from the ms and the
 hosts.

 Daan (double a!)

 On Fri, Nov 1, 2013 at 7:52 PM, Carlos Reategui create...@gmail.com
 wrote:
  Hi Dan,
  The host value is correct and I am able to connect on ports 22,80 and
 8080
  from the Xen hosts to that IP (same subnet).  Port 8787 is not listening
 on
  my MS.
 
  What can I look for in the DB?
 
  thanks
  Carlos
 
 
 
 
  On Fri, Nov 1, 2013 at 2:47 AM, Daan Hoogland 

Re: Can't destroy system VMs

2013-11-01 Thread Carlos Reategui
Hi Dan,
The host value is correct and I am able to connect on ports 22,80 and 8080
from the Xen hosts to that IP (same subnet).  Port 8787 is not listening on
my MS.

What can I look for in the DB?

thanks
Carlos




On Fri, Nov 1, 2013 at 2:47 AM, Daan Hoogland daan.hoogl...@gmail.comwrote:

 H Carlox,

 It still seems to me you have a communication problem albeit not at
 the network level. ( I agree this doesn't seem likely) But even so the
 ms can't reach the xen hosts. The next thing to look at would be
 configuration. I presume you did check the ip addresses for the hosts.
 check the value 'host' in the global vars and try to telnet it on port
 22 80 8080 8787 or what else from the xen hosts . I am sure the
 problem is hidden somewhere in the mysql database.

 regards,
 Daan

 On Fri, Nov 1, 2013 at 6:23 AM, Carlos Reategui car...@reategui.com
 wrote:
  Hi Dan,
  Finally getting around to trying this.  I am not able to start any vm
 (mine
  or system).
 
  I am in the process of trying Ahmad's suggestion of removing one of the
  hosts, reinstalling it and then adding it back in to see if that helps.
  However I am not able to put it in maintenance mode.  I am getting the
  following when I try:
 
 
  2013-10-31 15:19:59,120 DEBUG [cloud.api.ApiServlet]
 (catalina-exec-5:null)
  ===START===  172.30.40.135 -- GET  command=prepareHostForMa
 
 intenanceid=c2f3c416-f4ba-4f73-b414-6fd6efc734e3response=jsonsessionkey=Yjye5qmcLtUBW0%2FP4XCAcx2ZTAI%3D_=1383258001952
  2013-10-31 15:19:59,181 DEBUG [cloud.async.AsyncJobManagerImpl]
  (catalina-exec-5:null) submit async job-167, details: AsyncJobVO {id:16
  7, userId: 2, accountId: 2, sessionKey: null, instanceType: Host,
  instanceId: 1, cmd: org.apache.cloudstack.api.command.admin.host.Prep
  areForMaintenanceCmd, cmdOriginator: null, cmdInfo:
 
 {id:c2f3c416-f4ba-4f73-b414-6fd6efc734e3,response:json,sessionkey:Yjye5q
 
 mcLtUBW0/P4XCAcx2ZTAI\u003d,ctxUserId:2,_:1383258001952,ctxAccountId:2,ctxStartEventId:699},
  cmdVersion: 0, callbackTy
  pe: 0, callbackAddress: null, status: 0, processStatus: 0, resultCode: 0,
  result: null, initMsid: 233845174730253, completeMsid: null,
  lastUpdated: null, lastPolled: null, created: null}
  2013-10-31 15:19:59,185 DEBUG [cloud.async.AsyncJobManagerImpl]
  (Job-Executor-1:job-167) Executing org.apache.cloudstack.api.command.ad
  min.host.PrepareForMaintenanceCmd for job-167
  2013-10-31 15:19:59,188 DEBUG [cloud.api.ApiServlet]
 (catalina-exec-5:null)
  ===END===  172.30.40.135 -- GET  command=prepareHostForMain
 
 tenanceid=c2f3c416-f4ba-4f73-b414-6fd6efc734e3response=jsonsessionkey=Yjye5qmcLtUBW0%2FP4XCAcx2ZTAI%3D_=1383258001952
  2013-10-31 15:19:59,207 DEBUG [cloud.cluster.ClusterManagerImpl]
  (Job-Executor-1:job-167) Propagating agent change request event:AdminA
  skMaintenace to agent:1
  2013-10-31 15:19:59,208 DEBUG [cloud.cluster.ClusterManagerImpl]
  (Job-Executor-1:job-167) 233845174730253 - 233845174730255.1 [{Propa
 
 gateResourceEventCommand:{hostId:1,event:AdminAskMaintenace,contextMap:{},wait:0}}]
  2013-10-31 15:19:59,213 INFO  [cloud.cluster.ClusterServiceServletImpl]
  (Cluster-Worker-1:null) Setup cluster service servlet. service
  url: http://172.30.45.2:9090/clusterservice, request timeout: 300
 seconds
  2013-10-31 15:19:59,214 DEBUG [cloud.cluster.ClusterManagerImpl]
  (Cluster-Worker-1:null) Cluster PDU 233845174730253 - 233845174730255
  . agent: 1, pdu seq: 1, pdu ack seq: 0, json:
 
 [{PropagateResourceEventCommand:{hostId:1,event:AdminAskMaintenace,contextMap:{
  },wait:0}}]
  2013-10-31 15:19:59,343 DEBUG [cloud.cluster.ClusterManagerImpl]
  (Cluster-Worker-7:null) Dispatch -1, json: [{PropagateResourceEventC
 
 ommand:{hostId:1,event:AdminAskMaintenace,contextMap:{},wait:0}}]
  2013-10-31 15:19:59,347 DEBUG [cloud.cluster.ClusterManagerImpl]
  (Cluster-Worker-7:null) Intercepting command to propagate event AdminA
  skMaintenace for host 1
  2013-10-31 15:19:59,351 DEBUG [cloud.cluster.ClusterServiceServletImpl]
  (Cluster-Worker-1:null) POST http://172.30.45.2:9090/clusterser
  vice response :true, responding time: 80 ms
  2013-10-31 15:19:59,351 DEBUG [cloud.cluster.ClusterManagerImpl]
  (Cluster-Worker-1:null) Cluster PDU 233845174730253 - 233845174730255
   completed. time: 137ms. agent: 1, pdu seq: 1, pdu ack seq: 0, json:
  [{PropagateResourceEventCommand:{hostId:1,event:AdminAskMai
  ntenace,contextMap:{},wait:0}}]
  2013-10-31 15:19:59,356 DEBUG [agent.manager.ClusteredAgentAttache]
  (Cluster-Worker-7:null) Seq 1-102760483: Forwarding Seq 1-102760483
  :  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
  [{MaintainCommand:{wait:0}}] } to 233845174730255
  2013-10-31 15:19:59,360 DEBUG [agent.manager.ClusteredAgentAttache]
  (AgentManager-Handler-1:null) Seq 1-102760483: Forwarding Seq 1-102
  760483:  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
  [{MaintainCommand:{wait:0}}] } to 233845174730255
  2013-10-31 15:19:59,365 DEBUG 

Re: Can't destroy system VMs

2013-10-31 Thread Daan Hoogland
H Carlos,

Did you solve this problem yet?

You can try removing the cpvm and vr at the hypervisor and change
their state to 'Stopped' in the db.

On Tue, Oct 29, 2013 at 8:03 PM, Carlos Reategui car...@reategui.com wrote:
 In the last 15 minutes management log file it grew 500MB..

 Any ideas what I should look at to figure out what is happening?


 On Tue, Oct 29, 2013 at 11:53 AM, Carlos Reategui create...@gmail.comwrote:

 I am trying to recover my CloudStack installation (CS 4.1.1 on ubuntu + XS
 6.0.2) that I shutdown and cannot bring back to life.

 I tried to destroy the CPVM and the VR and all I am seeing in the logs is
 the following and they are filling up fast (management log is almost 1GB).


 2013-10-29 11:51:08,291 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-14:null) Seq 1-201919186: Forwarding Seq 1-201919186:
  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
 [{storage.DestroyCommand:{vmName:v-2-VM,volume:{id:2,name:ROOT-2,mountPoint:/export/primary,path:9aa8e81d-edb8-4284-a422-a89e9fd4237c,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,293 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-4:null) Seq 2-1168706613: Forwarding Seq
 2-1168706613:  { Cmd , MgmtId: 233845174730253, via: 2, Ver: v1, Flags:
 100111,
 [{storage.DestroyCommand:{vmName:r-5-VM,volume:{id:6,name:ROOT-5,mountPoint:/export/primary,path:56b215e0-bbb0-455f-9896-620ce22d28ad,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,293 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-2:null) Seq 1-201919186: Forwarding Seq 1-201919186:
  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
 [{storage.DestroyCommand:{vmName:v-2-VM,volume:{id:2,name:ROOT-2,mountPoint:/export/primary,path:9aa8e81d-edb8-4284-a422-a89e9fd4237c,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,294 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-8:null) Seq 2-1168706613: Forwarding Seq
 2-1168706613:  { Cmd , MgmtId: 233845174730253, via: 2, Ver: v1, Flags:
 100111,
 [{storage.DestroyCommand:{vmName:r-5-VM,volume:{id:6,name:ROOT-5,mountPoint:/export/primary,path:56b215e0-bbb0-455f-9896-620ce22d28ad,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,294 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-9:null) Seq 1-201919186: Forwarding Seq 1-201919186:
  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
 [{storage.DestroyCommand:{vmName:v-2-VM,volume:{id:2,name:ROOT-2,mountPoint:/export/primary,path:9aa8e81d-edb8-4284-a422-a89e9fd4237c,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,296 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-6:null) Seq 2-1168706613: Forwarding Seq
 2-1168706613:  { Cmd , MgmtId: 233845174730253, via: 2, Ver: v1, Flags:
 100111,
 [{storage.DestroyCommand:{vmName:r-5-VM,volume:{id:6,name:ROOT-5,mountPoint:/export/primary,path:56b215e0-bbb0-455f-9896-620ce22d28ad,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,296 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-15:null) Seq 1-201919186: Forwarding Seq 1-201919186:
  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
 [{storage.DestroyCommand:{vmName:v-2-VM,volume:{id:2,name:ROOT-2,mountPoint:/export/primary,path:9aa8e81d-edb8-4284-a422-a89e9fd4237c,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,297 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-13:null) Seq 2-1168706613: Forwarding Seq
 2-1168706613:  { Cmd , MgmtId: 233845174730253, via: 2, Ver: v1, Flags:
 100111,
 [{storage.DestroyCommand:{vmName:r-5-VM,volume:{id:6,name:ROOT-5,mountPoint:/export/primary,path:56b215e0-bbb0-455f-9896-620ce22d28ad,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,297 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-10:null) Seq 1-201919186: Forwarding Seq 1-201919186:
  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
 

Re: Can't destroy system VMs

2013-10-31 Thread Carlos Reategui
Hi Dan,

Unfortunately not.  Still trying to solve this.  See my last email to Ahmad
in the recover cs thread about next steps I am trying to figure out.

The CPVM and VR are not running.  They are currently in expunging state.

If I try to start an instance the MS is unable to communicate the start vm
commands to my XS hosts and then starts writing to the logs at an amazingly
high rate.  You can see one of my logs here:
http://reategui.com/cloudstack/management-server.log.new

thank you
Carlos





On Thu, Oct 31, 2013 at 1:50 PM, Daan Hoogland daan.hoogl...@gmail.comwrote:

 H Carlos,

 Did you solve this problem yet?

 You can try removing the cpvm and vr at the hypervisor and change
 their state to 'Stopped' in the db.

 On Tue, Oct 29, 2013 at 8:03 PM, Carlos Reategui car...@reategui.com
 wrote:
  In the last 15 minutes management log file it grew 500MB..
 
  Any ideas what I should look at to figure out what is happening?
 
 
  On Tue, Oct 29, 2013 at 11:53 AM, Carlos Reategui create...@gmail.com
 wrote:
 
  I am trying to recover my CloudStack installation (CS 4.1.1 on ubuntu +
 XS
  6.0.2) that I shutdown and cannot bring back to life.
 
  I tried to destroy the CPVM and the VR and all I am seeing in the logs
 is
  the following and they are filling up fast (management log is almost
 1GB).
 
 
  2013-10-29 11:51:08,291 DEBUG [agent.manager.ClusteredAgentAttache]
  (AgentManager-Handler-14:null) Seq 1-201919186: Forwarding Seq
 1-201919186:
   { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
 
 [{storage.DestroyCommand:{vmName:v-2-VM,volume:{id:2,name:ROOT-2,mountPoint:/export/primary,path:9aa8e81d-edb8-4284-a422-a89e9fd4237c,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
  } to 233845174730255
  2013-10-29 11:51:08,293 DEBUG [agent.manager.ClusteredAgentAttache]
  (AgentManager-Handler-4:null) Seq 2-1168706613: Forwarding Seq
  2-1168706613:  { Cmd , MgmtId: 233845174730253, via: 2, Ver: v1, Flags:
  100111,
 
 [{storage.DestroyCommand:{vmName:r-5-VM,volume:{id:6,name:ROOT-5,mountPoint:/export/primary,path:56b215e0-bbb0-455f-9896-620ce22d28ad,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
  } to 233845174730255
  2013-10-29 11:51:08,293 DEBUG [agent.manager.ClusteredAgentAttache]
  (AgentManager-Handler-2:null) Seq 1-201919186: Forwarding Seq
 1-201919186:
   { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
 
 [{storage.DestroyCommand:{vmName:v-2-VM,volume:{id:2,name:ROOT-2,mountPoint:/export/primary,path:9aa8e81d-edb8-4284-a422-a89e9fd4237c,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
  } to 233845174730255
  2013-10-29 11:51:08,294 DEBUG [agent.manager.ClusteredAgentAttache]
  (AgentManager-Handler-8:null) Seq 2-1168706613: Forwarding Seq
  2-1168706613:  { Cmd , MgmtId: 233845174730253, via: 2, Ver: v1, Flags:
  100111,
 
 [{storage.DestroyCommand:{vmName:r-5-VM,volume:{id:6,name:ROOT-5,mountPoint:/export/primary,path:56b215e0-bbb0-455f-9896-620ce22d28ad,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
  } to 233845174730255
  2013-10-29 11:51:08,294 DEBUG [agent.manager.ClusteredAgentAttache]
  (AgentManager-Handler-9:null) Seq 1-201919186: Forwarding Seq
 1-201919186:
   { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
 
 [{storage.DestroyCommand:{vmName:v-2-VM,volume:{id:2,name:ROOT-2,mountPoint:/export/primary,path:9aa8e81d-edb8-4284-a422-a89e9fd4237c,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
  } to 233845174730255
  2013-10-29 11:51:08,296 DEBUG [agent.manager.ClusteredAgentAttache]
  (AgentManager-Handler-6:null) Seq 2-1168706613: Forwarding Seq
  2-1168706613:  { Cmd , MgmtId: 233845174730253, via: 2, Ver: v1, Flags:
  100111,
 
 [{storage.DestroyCommand:{vmName:r-5-VM,volume:{id:6,name:ROOT-5,mountPoint:/export/primary,path:56b215e0-bbb0-455f-9896-620ce22d28ad,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
  } to 233845174730255
  2013-10-29 11:51:08,296 DEBUG [agent.manager.ClusteredAgentAttache]
  (AgentManager-Handler-15:null) Seq 1-201919186: Forwarding Seq
 1-201919186:
   { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
 
 [{storage.DestroyCommand:{vmName:v-2-VM,volume:{id:2,name:ROOT-2,mountPoint:/export/primary,path:9aa8e81d-edb8-4284-a422-a89e9fd4237c,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
  } to 233845174730255
  2013-10-29 11:51:08,297 DEBUG [agent.manager.ClusteredAgentAttache]
  

Re: Can't destroy system VMs

2013-10-31 Thread Carlos Reategui
Hi Dan,
Network is definitely not a problem.  iptables is disabled on MS and both
hosts.  I am able to ssh from MS to both hosts.  I am also able to connect
to port 80 from the MS.

I need to head out now and will try what you suggested in a couple hours.

thanks,
Carlos


On Thu, Oct 31, 2013 at 3:19 PM, Daan Hoogland daan.hoogl...@gmail.comwrote:

 Carlos,

 can the ms reach the xs hosts?
 if so try this, Stop the ms, set the state for the vm_instances for
 cpvm and vr from expunging to stopped. and restart.
 if not you need to troubleshoot your network connectivity. It seems
 this would not be the issue as you only stopped and started it. Did
 you

 Daan

 On Thu, Oct 31, 2013 at 10:19 PM, Carlos Reategui car...@reategui.com
 wrote:
  Hi Dan,
 
  Unfortunately not.  Still trying to solve this.  See my last email to
 Ahmad
  in the recover cs thread about next steps I am trying to figure out.
 
  The CPVM and VR are not running.  They are currently in expunging state.
 
  If I try to start an instance the MS is unable to communicate the start
 vm
  commands to my XS hosts and then starts writing to the logs at an
 amazingly
  high rate.  You can see one of my logs here:
  http://reategui.com/cloudstack/management-server.log.new
 
  thank you
  Carlos
 
 
 
 
 
  On Thu, Oct 31, 2013 at 1:50 PM, Daan Hoogland daan.hoogl...@gmail.com
  wrote:
 
  H Carlos,
 
  Did you solve this problem yet?
 
  You can try removing the cpvm and vr at the hypervisor and change
  their state to 'Stopped' in the db.
 
  On Tue, Oct 29, 2013 at 8:03 PM, Carlos Reategui car...@reategui.com
  wrote:
   In the last 15 minutes management log file it grew 500MB..
  
   Any ideas what I should look at to figure out what is happening?
  
  
   On Tue, Oct 29, 2013 at 11:53 AM, Carlos Reategui
   create...@gmail.comwrote:
  
   I am trying to recover my CloudStack installation (CS 4.1.1 on
 ubuntu +
   XS
   6.0.2) that I shutdown and cannot bring back to life.
  
   I tried to destroy the CPVM and the VR and all I am seeing in the
 logs
   is
   the following and they are filling up fast (management log is almost
   1GB).
  
  
   2013-10-29 11:51:08,291 DEBUG [agent.manager.ClusteredAgentAttache]
   (AgentManager-Handler-14:null) Seq 1-201919186: Forwarding Seq
   1-201919186:
{ Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
  
  
 [{storage.DestroyCommand:{vmName:v-2-VM,volume:{id:2,name:ROOT-2,mountPoint:/export/primary,path:9aa8e81d-edb8-4284-a422-a89e9fd4237c,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
   } to 233845174730255
   2013-10-29 11:51:08,293 DEBUG [agent.manager.ClusteredAgentAttache]
   (AgentManager-Handler-4:null) Seq 2-1168706613: Forwarding Seq
   2-1168706613:  { Cmd , MgmtId: 233845174730253, via: 2, Ver: v1,
 Flags:
   100111,
  
  
 [{storage.DestroyCommand:{vmName:r-5-VM,volume:{id:6,name:ROOT-5,mountPoint:/export/primary,path:56b215e0-bbb0-455f-9896-620ce22d28ad,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
   } to 233845174730255
   2013-10-29 11:51:08,293 DEBUG [agent.manager.ClusteredAgentAttache]
   (AgentManager-Handler-2:null) Seq 1-201919186: Forwarding Seq
   1-201919186:
{ Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
  
  
 [{storage.DestroyCommand:{vmName:v-2-VM,volume:{id:2,name:ROOT-2,mountPoint:/export/primary,path:9aa8e81d-edb8-4284-a422-a89e9fd4237c,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
   } to 233845174730255
   2013-10-29 11:51:08,294 DEBUG [agent.manager.ClusteredAgentAttache]
   (AgentManager-Handler-8:null) Seq 2-1168706613: Forwarding Seq
   2-1168706613:  { Cmd , MgmtId: 233845174730253, via: 2, Ver: v1,
 Flags:
   100111,
  
  
 [{storage.DestroyCommand:{vmName:r-5-VM,volume:{id:6,name:ROOT-5,mountPoint:/export/primary,path:56b215e0-bbb0-455f-9896-620ce22d28ad,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
   } to 233845174730255
   2013-10-29 11:51:08,294 DEBUG [agent.manager.ClusteredAgentAttache]
   (AgentManager-Handler-9:null) Seq 1-201919186: Forwarding Seq
   1-201919186:
{ Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
  
  
 [{storage.DestroyCommand:{vmName:v-2-VM,volume:{id:2,name:ROOT-2,mountPoint:/export/primary,path:9aa8e81d-edb8-4284-a422-a89e9fd4237c,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
   } to 233845174730255
   2013-10-29 11:51:08,296 DEBUG [agent.manager.ClusteredAgentAttache]
   (AgentManager-Handler-6:null) Seq 2-1168706613: Forwarding Seq
   2-1168706613:  { Cmd , MgmtId: 233845174730253, via: 2, Ver: v1,
 Flags:
   100111,
  
  
 

Re: Can't destroy system VMs

2013-10-29 Thread Carlos Reategui
In the last 15 minutes management log file it grew 500MB..

Any ideas what I should look at to figure out what is happening?


On Tue, Oct 29, 2013 at 11:53 AM, Carlos Reategui create...@gmail.comwrote:

 I am trying to recover my CloudStack installation (CS 4.1.1 on ubuntu + XS
 6.0.2) that I shutdown and cannot bring back to life.

 I tried to destroy the CPVM and the VR and all I am seeing in the logs is
 the following and they are filling up fast (management log is almost 1GB).


 2013-10-29 11:51:08,291 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-14:null) Seq 1-201919186: Forwarding Seq 1-201919186:
  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
 [{storage.DestroyCommand:{vmName:v-2-VM,volume:{id:2,name:ROOT-2,mountPoint:/export/primary,path:9aa8e81d-edb8-4284-a422-a89e9fd4237c,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,293 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-4:null) Seq 2-1168706613: Forwarding Seq
 2-1168706613:  { Cmd , MgmtId: 233845174730253, via: 2, Ver: v1, Flags:
 100111,
 [{storage.DestroyCommand:{vmName:r-5-VM,volume:{id:6,name:ROOT-5,mountPoint:/export/primary,path:56b215e0-bbb0-455f-9896-620ce22d28ad,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,293 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-2:null) Seq 1-201919186: Forwarding Seq 1-201919186:
  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
 [{storage.DestroyCommand:{vmName:v-2-VM,volume:{id:2,name:ROOT-2,mountPoint:/export/primary,path:9aa8e81d-edb8-4284-a422-a89e9fd4237c,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,294 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-8:null) Seq 2-1168706613: Forwarding Seq
 2-1168706613:  { Cmd , MgmtId: 233845174730253, via: 2, Ver: v1, Flags:
 100111,
 [{storage.DestroyCommand:{vmName:r-5-VM,volume:{id:6,name:ROOT-5,mountPoint:/export/primary,path:56b215e0-bbb0-455f-9896-620ce22d28ad,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,294 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-9:null) Seq 1-201919186: Forwarding Seq 1-201919186:
  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
 [{storage.DestroyCommand:{vmName:v-2-VM,volume:{id:2,name:ROOT-2,mountPoint:/export/primary,path:9aa8e81d-edb8-4284-a422-a89e9fd4237c,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,296 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-6:null) Seq 2-1168706613: Forwarding Seq
 2-1168706613:  { Cmd , MgmtId: 233845174730253, via: 2, Ver: v1, Flags:
 100111,
 [{storage.DestroyCommand:{vmName:r-5-VM,volume:{id:6,name:ROOT-5,mountPoint:/export/primary,path:56b215e0-bbb0-455f-9896-620ce22d28ad,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,296 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-15:null) Seq 1-201919186: Forwarding Seq 1-201919186:
  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,
 [{storage.DestroyCommand:{vmName:v-2-VM,volume:{id:2,name:ROOT-2,mountPoint:/export/primary,path:9aa8e81d-edb8-4284-a422-a89e9fd4237c,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,297 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-13:null) Seq 2-1168706613: Forwarding Seq
 2-1168706613:  { Cmd , MgmtId: 233845174730253, via: 2, Ver: v1, Flags:
 100111,
 [{storage.DestroyCommand:{vmName:r-5-VM,volume:{id:6,name:ROOT-5,mountPoint:/export/primary,path:56b215e0-bbb0-455f-9896-620ce22d28ad,size:2147483648,type:ROOT,storagePoolType:NetworkFilesystem,storagePoolUuid:17b0a8a5-2376-3d11-b60e-31eebeafb217,deviceId:0},wait:0}}]
 } to 233845174730255
 2013-10-29 11:51:08,297 DEBUG [agent.manager.ClusteredAgentAttache]
 (AgentManager-Handler-10:null) Seq 1-201919186: Forwarding Seq 1-201919186:
  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags: 100111,