Sorry Carlos,

family and $dayjob kept me from looking at this anymore. Glad you got
it working.

Daan

On Wed, Nov 6, 2013 at 8:08 PM, Carlos Reategui <car...@reategui.com> wrote:
> Since I didn't hear back I bit the bullet and decided to make the DB
> changes and I am back up and running.  Woohoo.  I'll post the specifics in
> the other thread I started about the mshost id.
>
>
> On Tue, Nov 5, 2013 at 11:36 AM, Carlos Reategui <car...@reategui.com>wrote:
>
>> Hi Daan,
>>
>> The mshost thing I mentioned previously has been bugging me and as I look
>> through the db I think I have a problem here.  I am guessing I have a
>> botched 4.1.0 to 4.1.1 upgrade that I did not notice when I shut things
>> down.
>>
>> The mshost table is showing 2 management servers that are supposedly on
>> the same machine listening on the same port which is not really possible.
>>  So really only one of the 2 is up (I believe that would be the 4.1.1
>> version). The logs do seem to indicate something along those lines where
>> the 2 management servers are failing to communicate with each other.
>>
>> If I look at the host table I see my 2 hosts but they say they are on
>> version 4.1.0 and are pointing to management server id of the 4.1.0 line in
>> the mshost table.
>>
>> I am wondering if this is the problem and why their is no communication to
>> my hosts from the management server.
>>
>> Anyone know the schema well enough to help me update some stuff and
>> recover my installation?  Would an upgrade to 4.2 possibly cleanup my DB?
>>
>> thanks,
>> Carlos
>>
>>
>>
>>
>> On Tue, Nov 5, 2013 at 10:59 AM, Carlos Reategui <car...@reategui.com>wrote:
>>
>>> Hi Daan,
>>>
>>> I have been looking through my logs and can't figure out why the ssvm
>>> does not want to start.
>>>
>>> Here is a link to my latest logs:
>>> http://reategui.com/cloudstack/management-server-131105.log
>>>
>>> Maybe someone else can spot something.
>>>
>>> There is nothing on either the XenServer hosts logs.
>>>
>>> thanks,
>>> Carlos
>>>
>>>
>>> On Tue, Nov 5, 2013 at 12:55 AM, Daan Hoogland 
>>> <daan.hoogl...@gmail.com>wrote:
>>>
>>>> Carlos,
>>>>
>>>> It makes no sense trying to start the vms if not the ssvm is started
>>>> first.
>>>>
>>>> I would say stop the system, clear the logs and start again. Then
>>>> analyse the logs to see what is wrong with the secondary storage vm.
>>>>
>>>> On Tue, Nov 5, 2013 at 2:08 AM, Carlos Reategui <car...@reategui.com>
>>>> wrote:
>>>> > Hi Daan,
>>>> > There is no trace of CS communicating with either of my XS hosts in the
>>>> > SMLog file or the xensource.log files when I try to start one of my
>>>> stopped
>>>> > instances.
>>>> >
>>>> > thanks,
>>>> > Carlos
>>>> >
>>>> >
>>>> >
>>>> > On Sat, Nov 2, 2013 at 4:50 AM, Daan Hoogland <daan.hoogl...@gmail.com
>>>> >wrote:
>>>> >
>>>> >> H Carlos,
>>>> >>
>>>> >> I wouldn't know about the mshost and mshost_peer but they seem
>>>> >> alright. The expunging ssvm are a real problem. I think Ahmed
>>>> >> mentioned that as well. You are having problems getting your ssvm in
>>>> >> the air. All other things are consequential. This is the first thing
>>>> >> to look at. Check the logs and the SMLog on the hypervisor to see what
>>>> >> happens when cs tries to start a ssvm. By your db you can see that it
>>>> >> tried to start it several times. Those are the expunging vm_instance
>>>> >> entries.
>>>> >>
>>>> >> regards,
>>>> >> Daan
>>>> >>
>>>> >> On Fri, Nov 1, 2013 at 9:10 PM, Carlos Reategui <car...@reategui.com>
>>>> >> wrote:
>>>> >> > Hi Daan  :)
>>>> >> >
>>>> >> > Primary and Secondary storage are on the MS.  Primary is on a
>>>> different
>>>> >> > subnet and it is currently mounted (and visible as an NFS SR from
>>>> >> > XenCenter) and working fine from the xen machines.  I have also
>>>> been able
>>>> >> > to mount Secondary without a problem.
>>>> >> >
>>>> >> > Here are my current system VMs:
>>>> >> > mysql> select
>>>> >> > id,name,instance_name,state,host_id,type,vm_type,desired_state from
>>>> >> > vm_instance where type!='User';
>>>> >> >
>>>> >>
>>>> +----+---------+---------------+-----------+---------+--------------------+--------------------+---------------+
>>>> >> > | id | name    | instance_name | state     | host_id | type
>>>> >>   |
>>>> >> > vm_type            | desired_state |
>>>> >> >
>>>> >>
>>>> +----+---------+---------------+-----------+---------+--------------------+--------------------+---------------+
>>>> >> > |  2 | v-2-VM  | v-2-VM        | Stopped   |       1 | ConsoleProxy
>>>> >>   |
>>>> >> > ConsoleProxy       | NULL          |
>>>> >> > |  5 | r-5-VM  | r-5-VM        | Stopped   |       2 | DomainRouter
>>>> >>   |
>>>> >> > DomainRouter       | NULL          |
>>>> >> > |  1 | s-1-VM  | s-1-VM        | Expunging |    NULL |
>>>> >> SecondaryStorageVm |
>>>> >> > SecondaryStorageVm | NULL          |
>>>> >> > |  3 | s-3-VM  | s-3-VM        | Expunging |    NULL |
>>>> >> SecondaryStorageVm |
>>>> >> > SecondaryStorageVm | NULL          |
>>>> >> > | 52 | s-52-VM | s-52-VM       | Expunging |    NULL |
>>>> >> SecondaryStorageVm |
>>>> >> > SecondaryStorageVm | NULL          |
>>>> >> > | 53 | s-53-VM | s-53-VM       | Expunging |    NULL |
>>>> >> SecondaryStorageVm |
>>>> >> > SecondaryStorageVm | NULL          |
>>>> >> > | 54 | s-54-VM | s-54-VM       | Expunging |    NULL |
>>>> >> SecondaryStorageVm |
>>>> >> > SecondaryStorageVm | NULL          |
>>>> >> > | 55 | s-55-VM | s-55-VM       | Expunging |    NULL |
>>>> >> SecondaryStorageVm |
>>>> >> > SecondaryStorageVm | NULL          |
>>>> >> > | 56 | s-56-VM | s-56-VM       | Expunging |    NULL |
>>>> >> SecondaryStorageVm |
>>>> >> > SecondaryStorageVm | NULL          |
>>>> >> >
>>>> >>
>>>> +----+---------+---------------+-----------+---------+--------------------+--------------------+---------------+
>>>> >> >
>>>> >> >
>>>> >> > I have been looking through the DB to see if anything sticks out.  I
>>>> >> notice
>>>> >> > that the mshost and mshost_peer table have 2 entries but they are
>>>> the
>>>> >> same
>>>> >> > server.  Is that ok?  Looks like one is from the install of 4.1.0
>>>> an the
>>>> >> > other from after I upgraded to 4.1.1:
>>>> >> >
>>>> >> > mysql> select * from mshost;
>>>> >> >
>>>> >>
>>>> +----+-----------------+---------------+------------+-------+---------+-------------+--------------+---------------------+---------+-------------+
>>>> >> > | id | msid            | runid         | name       | state |
>>>> version |
>>>> >> > service_ip  | service_port | last_update         | removed |
>>>> alert_count
>>>> >> |
>>>> >> >
>>>> >>
>>>> +----+-----------------+---------------+------------+-------+---------+-------------+--------------+---------------------+---------+-------------+
>>>> >> > |  1 | 233845174730255 | 1375383759777 | isbustor01 | Up    | 4.1.0
>>>>   |
>>>> >> > 172.30.45.2 |         9090 | 2013-10-24 21:13:10 | NULL    |
>>>>     0
>>>> >> |
>>>> >> > |  2 | 233845174730253 | 1383331508626 | isbustor01 | Up    | 4.1.1
>>>>   |
>>>> >> > 172.30.45.2 |         9090 | 2013-11-01 20:00:04 | NULL    |
>>>>     0
>>>> >> |
>>>> >> >
>>>> >>
>>>> +----+-----------------+---------------+------------+-------+---------+-------------+--------------+---------------------+---------+-------------+
>>>> >> > 2 rows in set (0.00 sec)
>>>> >> >
>>>> >> > mysql> select * from mshost_peer;
>>>> >> >
>>>> >>
>>>> +----+--------------+-------------+---------------+------------+---------------------+
>>>> >> > | id | owner_mshost | peer_mshost | peer_runid    | peer_state |
>>>> >> > last_update         |
>>>> >> >
>>>> >>
>>>> +----+--------------+-------------+---------------+------------+---------------------+
>>>> >> > | 17 |            1 |           1 | 1375383759777 | Up         |
>>>> >> 2013-08-01
>>>> >> > 19:02:49 |
>>>> >> > | 34 |            2 |           2 | 1383331508626 | Up         |
>>>> >> 2013-11-01
>>>> >> > 18:45:19 |
>>>> >> >
>>>> >>
>>>> +----+--------------+-------------+---------------+------------+---------------------+
>>>> >> >
>>>> >> >
>>>> >> > thanks,
>>>> >> > Carlos
>>>> >> >
>>>> >> >
>>>> >> >
>>>> >> > On Fri, Nov 1, 2013 at 12:17 PM, Daan Hoogland <
>>>> daan.hoogl...@gmail.com
>>>> >> >wrote:
>>>> >> >
>>>> >> >> h Carlos,
>>>> >> >>
>>>> >> >> the states of virtual machines, especially the entries for the
>>>> ssvm.
>>>> >> >> if it is not Stopped. set it to stopped and check if your storage,
>>>> >> >> both primary and secondary, are reachable botjh from the ms and the
>>>> >> >> hosts.
>>>> >> >>
>>>> >> >> Daan (double a!)
>>>> >> >>
>>>> >> >> On Fri, Nov 1, 2013 at 7:52 PM, Carlos Reategui <
>>>> create...@gmail.com>
>>>> >> >> wrote:
>>>> >> >> > Hi Dan,
>>>> >> >> > The host value is correct and I am able to connect on ports
>>>> 22,80 and
>>>> >> >> 8080
>>>> >> >> > from the Xen hosts to that IP (same subnet).  Port 8787 is not
>>>> >> listening
>>>> >> >> on
>>>> >> >> > my MS.
>>>> >> >> >
>>>> >> >> > What can I look for in the DB?
>>>> >> >> >
>>>> >> >> > thanks
>>>> >> >> > Carlos
>>>> >> >> >
>>>> >> >> >
>>>> >> >> >
>>>> >> >> >
>>>> >> >> > On Fri, Nov 1, 2013 at 2:47 AM, Daan Hoogland <
>>>> >> daan.hoogl...@gmail.com>
>>>> >> >> > wrote:
>>>> >> >> >>
>>>> >> >> >> H Carlox,
>>>> >> >> >>
>>>> >> >> >> It still seems to me you have a communication problem albeit
>>>> not at
>>>> >> >> >> the network level. ( I agree this doesn't seem likely) But even
>>>> so
>>>> >> the
>>>> >> >> >> ms can't reach the xen hosts. The next thing to look at would be
>>>> >> >> >> configuration. I presume you did check the ip addresses for the
>>>> >> hosts.
>>>> >> >> >> check the value 'host' in the global vars and try to telnet it
>>>> on
>>>> >> port
>>>> >> >> >> 22 80 8080 8787 or what else from the xen hosts . I am sure the
>>>> >> >> >> problem is hidden somewhere in the mysql database.
>>>> >> >> >>
>>>> >> >> >> regards,
>>>> >> >> >> Daan
>>>> >> >> >>
>>>> >> >> >> On Fri, Nov 1, 2013 at 6:23 AM, Carlos Reategui <
>>>> car...@reategui.com
>>>> >> >
>>>> >> >> >> wrote:
>>>> >> >> >> > Hi Dan,
>>>> >> >> >> > Finally getting around to trying this.  I am not able to
>>>> start any
>>>> >> vm
>>>> >> >> >> > (mine
>>>> >> >> >> > or system).
>>>> >> >> >> >
>>>> >> >> >> > I am in the process of trying Ahmad's suggestion of removing
>>>> one of
>>>> >> >> the
>>>> >> >> >> > hosts, reinstalling it and then adding it back in to see if
>>>> that
>>>> >> >> helps.
>>>> >> >> >> > However I am not able to put it in maintenance mode.  I am
>>>> getting
>>>> >> the
>>>> >> >> >> > following when I try:
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >> >> > 2013-10-31 15:19:59,120 DEBUG [cloud.api.ApiServlet]
>>>> >> >> >> > (catalina-exec-5:null)
>>>> >> >> >> > ===START===  172.30.40.135 -- GET  command=prepareHostForMa
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >>
>>>> >>
>>>> intenance&id=c2f3c416-f4ba-4f73-b414-6fd6efc734e3&response=json&sessionkey=Yjye5qmcLtUBW0%2FP4XCAcx2ZTAI%3D&_=1383258001952
>>>> >> >> >> > 2013-10-31 15:19:59,181 DEBUG
>>>> [cloud.async.AsyncJobManagerImpl]
>>>> >> >> >> > (catalina-exec-5:null) submit async job-167, details:
>>>> AsyncJobVO
>>>> >> >> {id:16
>>>> >> >> >> > 7, userId: 2, accountId: 2, sessionKey: null, instanceType:
>>>> Host,
>>>> >> >> >> > instanceId: 1, cmd:
>>>> >> org.apache.cloudstack.api.command.admin.host.Prep
>>>> >> >> >> > areForMaintenanceCmd, cmdOriginator: null, cmdInfo:
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >>
>>>> >>
>>>> {"id":"c2f3c416-f4ba-4f73-b414-6fd6efc734e3","response":"json","sessionkey":"Yjye5q
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >>
>>>> >>
>>>> mcLtUBW0/P4XCAcx2ZTAI\u003d","ctxUserId":"2","_":"1383258001952","ctxAccountId":"2","ctxStartEventId":"699"},
>>>> >> >> >> > cmdVersion: 0, callbackTy
>>>> >> >> >> > pe: 0, callbackAddress: null, status: 0, processStatus: 0,
>>>> >> resultCode:
>>>> >> >> >> > 0,
>>>> >> >> >> > result: null, initMsid: 233845174730253, completeMsid: null,
>>>> >> >> >> > lastUpdated: null, lastPolled: null, created: null}
>>>> >> >> >> > 2013-10-31 15:19:59,185 DEBUG
>>>> [cloud.async.AsyncJobManagerImpl]
>>>> >> >> >> > (Job-Executor-1:job-167) Executing
>>>> >> >> org.apache.cloudstack.api.command.ad
>>>> >> >> >> > min.host.PrepareForMaintenanceCmd for job-167
>>>> >> >> >> > 2013-10-31 15:19:59,188 DEBUG [cloud.api.ApiServlet]
>>>> >> >> >> > (catalina-exec-5:null)
>>>> >> >> >> > ===END===  172.30.40.135 -- GET  command=prepareHostForMain
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >>
>>>> >>
>>>> tenance&id=c2f3c416-f4ba-4f73-b414-6fd6efc734e3&response=json&sessionkey=Yjye5qmcLtUBW0%2FP4XCAcx2ZTAI%3D&_=1383258001952
>>>> >> >> >> > 2013-10-31 15:19:59,207 DEBUG
>>>> [cloud.cluster.ClusterManagerImpl]
>>>> >> >> >> > (Job-Executor-1:job-167) Propagating agent change request
>>>> >> event:AdminA
>>>> >> >> >> > skMaintenace to agent:1
>>>> >> >> >> > 2013-10-31 15:19:59,208 DEBUG
>>>> [cloud.cluster.ClusterManagerImpl]
>>>> >> >> >> > (Job-Executor-1:job-167) 233845174730253 -> 233845174730255.1
>>>> >> [{"Propa
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >>
>>>> >>
>>>> gateResourceEventCommand":{"hostId":1,"event":"AdminAskMaintenace","contextMap":{},"wait":0}}]
>>>> >> >> >> > 2013-10-31 15:19:59,213 INFO
>>>> >> >>  [cloud.cluster.ClusterServiceServletImpl]
>>>> >> >> >> > (Cluster-Worker-1:null) Setup cluster service servlet. service
>>>> >> >> >> > url: http://172.30.45.2:9090/clusterservice, request
>>>> timeout: 300
>>>> >> >> >> > seconds
>>>> >> >> >> > 2013-10-31 15:19:59,214 DEBUG
>>>> [cloud.cluster.ClusterManagerImpl]
>>>> >> >> >> > (Cluster-Worker-1:null) Cluster PDU 233845174730253 ->
>>>> >> 233845174730255
>>>> >> >> >> > . agent: 1, pdu seq: 1, pdu ack seq: 0, json:
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >>
>>>> >>
>>>> [{"PropagateResourceEventCommand":{"hostId":1,"event":"AdminAskMaintenace","contextMap":{
>>>> >> >> >> > },"wait":0}}]
>>>> >> >> >> > 2013-10-31 15:19:59,343 DEBUG
>>>> [cloud.cluster.ClusterManagerImpl]
>>>> >> >> >> > (Cluster-Worker-7:null) Dispatch ->1, json:
>>>> >> [{"PropagateResourceEventC
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >>
>>>> >>
>>>> ommand":{"hostId":1,"event":"AdminAskMaintenace","contextMap":{},"wait":0}}]
>>>> >> >> >> > 2013-10-31 15:19:59,347 DEBUG
>>>> [cloud.cluster.ClusterManagerImpl]
>>>> >> >> >> > (Cluster-Worker-7:null) Intercepting command to propagate
>>>> event
>>>> >> AdminA
>>>> >> >> >> > skMaintenace for host 1
>>>> >> >> >> > 2013-10-31 15:19:59,351 DEBUG
>>>> >> >> [cloud.cluster.ClusterServiceServletImpl]
>>>> >> >> >> > (Cluster-Worker-1:null) POST
>>>> http://172.30.45.2:9090/clusterser
>>>> >> >> >> > vice response :true, responding time: 80 ms
>>>> >> >> >> > 2013-10-31 15:19:59,351 DEBUG
>>>> [cloud.cluster.ClusterManagerImpl]
>>>> >> >> >> > (Cluster-Worker-1:null) Cluster PDU 233845174730253 ->
>>>> >> 233845174730255
>>>> >> >> >> >  completed. time: 137ms. agent: 1, pdu seq: 1, pdu ack seq: 0,
>>>> >> json:
>>>> >> >> >> >
>>>> [{"PropagateResourceEventCommand":{"hostId":1,"event":"AdminAskMai
>>>> >> >> >> > ntenace","contextMap":{},"wait":0}}]
>>>> >> >> >> > 2013-10-31 15:19:59,356 DEBUG
>>>> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> > (Cluster-Worker-7:null) Seq 1-102760483: Forwarding Seq
>>>> 1-102760483
>>>> >> >> >> > :  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1, Flags:
>>>> 100111,
>>>> >> >> >> > [{"MaintainCommand":{"wait":0}}] } to 233845174730255
>>>> >> >> >> > 2013-10-31 15:19:59,360 DEBUG
>>>> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> > (AgentManager-Handler-1:null) Seq 1-102760483: Forwarding Seq
>>>> 1-102
>>>> >> >> >> > 760483:  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1,
>>>> Flags:
>>>> >> >> >> > 100111,
>>>> >> >> >> > [{"MaintainCommand":{"wait":0}}] } to 233845174730255
>>>> >> >> >> > 2013-10-31 15:19:59,365 DEBUG
>>>> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> > (AgentManager-Handler-13:null) Seq 1-102760483: Forwarding
>>>> Seq 1-10
>>>> >> >> >> > 2760483:  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1,
>>>> Flags:
>>>> >> >> >> > 100111,
>>>> >> >> >> > [{"MaintainCommand":{"wait":0}}] } to 233845174730255
>>>> >> >> >> > 2013-10-31 15:19:59,369 DEBUG
>>>> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> > (AgentManager-Handler-15:null) Seq 1-102760483: Forwarding
>>>> Seq 1-10
>>>> >> >> >> > 2760483:  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1,
>>>> Flags:
>>>> >> >> >> > 100111,
>>>> >> >> >> > [{"MaintainCommand":{"wait":0}}] } to 233845174730255
>>>> >> >> >> > 2013-10-31 15:19:59,372 DEBUG
>>>> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> > (AgentManager-Handler-10:null) Seq 1-102760483: Forwarding
>>>> Seq 1-10
>>>> >> >> >> > 2760483:  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1,
>>>> Flags:
>>>> >> >> >> > 100111,
>>>> >> >> >> > [{"MaintainCommand":{"wait":0}}] } to 233845174730255
>>>> >> >> >> > 2013-10-31 15:19:59,376 DEBUG
>>>> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> > (AgentManager-Handler-11:null) Seq 1-102760483: Forwarding
>>>> Seq 1-10
>>>> >> >> >> > 2760483:  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1,
>>>> Flags:
>>>> >> >> >> > 100111,
>>>> >> >> >> > [{"MaintainCommand":{"wait":0}}] } to 233845174730255
>>>> >> >> >> >
>>>> >> >> >> > These MaintainCommand lines are repeated endlessly and very
>>>> fast
>>>> >> (1GB
>>>> >> >> of
>>>> >> >> >> > logs in about 20 mins) so I currently have the MS stopped.
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >> >> >
>>>> >> >> >> > On Thu, Oct 31, 2013 at 3:43 PM, Carlos Reategui <
>>>> >> car...@reategui.com
>>>> >> >> >
>>>> >> >> >> > wrote:
>>>> >> >> >> >>
>>>> >> >> >> >> Hi Dan,
>>>> >> >> >> >> Network is definitely not a problem.  iptables is disabled
>>>> on MS
>>>> >> and
>>>> >> >> >> >> both
>>>> >> >> >> >> hosts.  I am able to ssh from MS to both hosts.  I am also
>>>> able to
>>>> >> >> >> >> connect
>>>> >> >> >> >> to port 80 from the MS.
>>>> >> >> >> >>
>>>> >> >> >> >> I need to head out now and will try what you suggested in a
>>>> couple
>>>> >> >> >> >> hours.
>>>> >> >> >> >>
>>>> >> >> >> >> thanks,
>>>> >> >> >> >> Carlos
>>>> >> >> >> >>
>>>> >> >> >> >>
>>>> >> >> >> >> On Thu, Oct 31, 2013 at 3:19 PM, Daan Hoogland
>>>> >> >> >> >> <daan.hoogl...@gmail.com>
>>>> >> >> >> >> wrote:
>>>> >> >> >> >>>
>>>> >> >> >> >>> Carlos,
>>>> >> >> >> >>>
>>>> >> >> >> >>> can the ms reach the xs hosts?
>>>> >> >> >> >>> if so try this, Stop the ms, set the state for the
>>>> vm_instances
>>>> >> for
>>>> >> >> >> >>> cpvm and vr from expunging to stopped. and restart.
>>>> >> >> >> >>> if not you need to troubleshoot your network connectivity.
>>>> It
>>>> >> seems
>>>> >> >> >> >>> this would not be the issue as you only stopped and started
>>>> it.
>>>> >> Did
>>>> >> >> >> >>> you
>>>> >> >> >> >>>
>>>> >> >> >> >>> Daan
>>>> >> >> >> >>>
>>>> >> >> >> >>> On Thu, Oct 31, 2013 at 10:19 PM, Carlos Reategui
>>>> >> >> >> >>> <car...@reategui.com>
>>>> >> >> >> >>> wrote:
>>>> >> >> >> >>> > Hi Dan,
>>>> >> >> >> >>> >
>>>> >> >> >> >>> > Unfortunately not.  Still trying to solve this.  See my
>>>> last
>>>> >> email
>>>> >> >> >> >>> > to
>>>> >> >> >> >>> > Ahmad
>>>> >> >> >> >>> > in the recover cs thread about next steps I am trying to
>>>> figure
>>>> >> >> out.
>>>> >> >> >> >>> >
>>>> >> >> >> >>> > The CPVM and VR are not running.  They are currently in
>>>> >> expunging
>>>> >> >> >> >>> > state.
>>>> >> >> >> >>> >
>>>> >> >> >> >>> > If I try to start an instance the MS is unable to
>>>> communicate
>>>> >> the
>>>> >> >> >> >>> > start
>>>> >> >> >> >>> > vm
>>>> >> >> >> >>> > commands to my XS hosts and then starts writing to the
>>>> logs at
>>>> >> an
>>>> >> >> >> >>> > amazingly
>>>> >> >> >> >>> > high rate.  You can see one of my logs here:
>>>> >> >> >> >>> > http://reategui.com/cloudstack/management-server.log.new
>>>> >> >> >> >>> >
>>>> >> >> >> >>> > thank you
>>>> >> >> >> >>> > Carlos
>>>> >> >> >> >>> >
>>>> >> >> >> >>> >
>>>> >> >> >> >>> >
>>>> >> >> >> >>> >
>>>> >> >> >> >>> >
>>>> >> >> >> >>> > On Thu, Oct 31, 2013 at 1:50 PM, Daan Hoogland
>>>> >> >> >> >>> > <daan.hoogl...@gmail.com>
>>>> >> >> >> >>> > wrote:
>>>> >> >> >> >>> >>
>>>> >> >> >> >>> >> H Carlos,
>>>> >> >> >> >>> >>
>>>> >> >> >> >>> >> Did you solve this problem yet?
>>>> >> >> >> >>> >>
>>>> >> >> >> >>> >> You can try removing the cpvm and vr at the hypervisor
>>>> and
>>>> >> change
>>>> >> >> >> >>> >> their state to 'Stopped' in the db.
>>>> >> >> >> >>> >>
>>>> >> >> >> >>> >> On Tue, Oct 29, 2013 at 8:03 PM, Carlos Reategui
>>>> >> >> >> >>> >> <car...@reategui.com>
>>>> >> >> >> >>> >> wrote:
>>>> >> >> >> >>> >> > In the last 15 minutes management log file it grew
>>>> 500MB..
>>>> >> >> >> >>> >> >
>>>> >> >> >> >>> >> > Any ideas what I should look at to figure out what is
>>>> >> >> happening?
>>>> >> >> >> >>> >> >
>>>> >> >> >> >>> >> >
>>>> >> >> >> >>> >> > On Tue, Oct 29, 2013 at 11:53 AM, Carlos Reategui
>>>> >> >> >> >>> >> > <create...@gmail.com>wrote:
>>>> >> >> >> >>> >> >
>>>> >> >> >> >>> >> >> I am trying to recover my CloudStack installation (CS
>>>> >> 4.1.1 on
>>>> >> >> >> >>> >> >> ubuntu +
>>>> >> >> >> >>> >> >> XS
>>>> >> >> >> >>> >> >> 6.0.2) that I shutdown and cannot bring back to life.
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >> I tried to destroy the CPVM and the VR and all I am
>>>> seeing
>>>> >> in
>>>> >> >> >> >>> >> >> the
>>>> >> >> >> >>> >> >> logs
>>>> >> >> >> >>> >> >> is
>>>> >> >> >> >>> >> >> the following and they are filling up fast
>>>> (management log
>>>> >> is
>>>> >> >> >> >>> >> >> almost
>>>> >> >> >> >>> >> >> 1GB).
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >> 2013-10-29 11:51:08,291 DEBUG
>>>> >> >> >> >>> >> >> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> >>> >> >> (AgentManager-Handler-14:null) Seq 1-201919186:
>>>> Forwarding
>>>> >> Seq
>>>> >> >> >> >>> >> >> 1-201919186:
>>>> >> >> >> >>> >> >>  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1,
>>>> Flags:
>>>> >> >> >> >>> >> >> 100111,
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >>
>>>> >>
>>>> [{"storage.DestroyCommand":{"vmName":"v-2-VM","volume":{"id":2,"name":"ROOT-2","mountPoint":"/export/primary","path":"9aa8e81d-edb8-4284-a422-a89e9fd4237c","size":2147483648,"type":"ROOT","storagePoolType":"NetworkFilesystem","storagePoolUuid":"17b0a8a5-2376-3d11-b60e-31eebeafb217","deviceId":0},"wait":0}}]
>>>> >> >> >> >>> >> >> } to 233845174730255
>>>> >> >> >> >>> >> >> 2013-10-29 11:51:08,293 DEBUG
>>>> >> >> >> >>> >> >> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> >>> >> >> (AgentManager-Handler-4:null) Seq 2-1168706613:
>>>> Forwarding
>>>> >> Seq
>>>> >> >> >> >>> >> >> 2-1168706613:  { Cmd , MgmtId: 233845174730253, via:
>>>> 2,
>>>> >> Ver:
>>>> >> >> v1,
>>>> >> >> >> >>> >> >> Flags:
>>>> >> >> >> >>> >> >> 100111,
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >>
>>>> >>
>>>> [{"storage.DestroyCommand":{"vmName":"r-5-VM","volume":{"id":6,"name":"ROOT-5","mountPoint":"/export/primary","path":"56b215e0-bbb0-455f-9896-620ce22d28ad","size":2147483648,"type":"ROOT","storagePoolType":"NetworkFilesystem","storagePoolUuid":"17b0a8a5-2376-3d11-b60e-31eebeafb217","deviceId":0},"wait":0}}]
>>>> >> >> >> >>> >> >> } to 233845174730255
>>>> >> >> >> >>> >> >> 2013-10-29 11:51:08,293 DEBUG
>>>> >> >> >> >>> >> >> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> >>> >> >> (AgentManager-Handler-2:null) Seq 1-201919186:
>>>> Forwarding
>>>> >> Seq
>>>> >> >> >> >>> >> >> 1-201919186:
>>>> >> >> >> >>> >> >>  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1,
>>>> Flags:
>>>> >> >> >> >>> >> >> 100111,
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >>
>>>> >>
>>>> [{"storage.DestroyCommand":{"vmName":"v-2-VM","volume":{"id":2,"name":"ROOT-2","mountPoint":"/export/primary","path":"9aa8e81d-edb8-4284-a422-a89e9fd4237c","size":2147483648,"type":"ROOT","storagePoolType":"NetworkFilesystem","storagePoolUuid":"17b0a8a5-2376-3d11-b60e-31eebeafb217","deviceId":0},"wait":0}}]
>>>> >> >> >> >>> >> >> } to 233845174730255
>>>> >> >> >> >>> >> >> 2013-10-29 11:51:08,294 DEBUG
>>>> >> >> >> >>> >> >> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> >>> >> >> (AgentManager-Handler-8:null) Seq 2-1168706613:
>>>> Forwarding
>>>> >> Seq
>>>> >> >> >> >>> >> >> 2-1168706613:  { Cmd , MgmtId: 233845174730253, via:
>>>> 2,
>>>> >> Ver:
>>>> >> >> v1,
>>>> >> >> >> >>> >> >> Flags:
>>>> >> >> >> >>> >> >> 100111,
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >>
>>>> >>
>>>> [{"storage.DestroyCommand":{"vmName":"r-5-VM","volume":{"id":6,"name":"ROOT-5","mountPoint":"/export/primary","path":"56b215e0-bbb0-455f-9896-620ce22d28ad","size":2147483648,"type":"ROOT","storagePoolType":"NetworkFilesystem","storagePoolUuid":"17b0a8a5-2376-3d11-b60e-31eebeafb217","deviceId":0},"wait":0}}]
>>>> >> >> >> >>> >> >> } to 233845174730255
>>>> >> >> >> >>> >> >> 2013-10-29 11:51:08,294 DEBUG
>>>> >> >> >> >>> >> >> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> >>> >> >> (AgentManager-Handler-9:null) Seq 1-201919186:
>>>> Forwarding
>>>> >> Seq
>>>> >> >> >> >>> >> >> 1-201919186:
>>>> >> >> >> >>> >> >>  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1,
>>>> Flags:
>>>> >> >> >> >>> >> >> 100111,
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >>
>>>> >>
>>>> [{"storage.DestroyCommand":{"vmName":"v-2-VM","volume":{"id":2,"name":"ROOT-2","mountPoint":"/export/primary","path":"9aa8e81d-edb8-4284-a422-a89e9fd4237c","size":2147483648,"type":"ROOT","storagePoolType":"NetworkFilesystem","storagePoolUuid":"17b0a8a5-2376-3d11-b60e-31eebeafb217","deviceId":0},"wait":0}}]
>>>> >> >> >> >>> >> >> } to 233845174730255
>>>> >> >> >> >>> >> >> 2013-10-29 11:51:08,296 DEBUG
>>>> >> >> >> >>> >> >> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> >>> >> >> (AgentManager-Handler-6:null) Seq 2-1168706613:
>>>> Forwarding
>>>> >> Seq
>>>> >> >> >> >>> >> >> 2-1168706613:  { Cmd , MgmtId: 233845174730253, via:
>>>> 2,
>>>> >> Ver:
>>>> >> >> v1,
>>>> >> >> >> >>> >> >> Flags:
>>>> >> >> >> >>> >> >> 100111,
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >>
>>>> >>
>>>> [{"storage.DestroyCommand":{"vmName":"r-5-VM","volume":{"id":6,"name":"ROOT-5","mountPoint":"/export/primary","path":"56b215e0-bbb0-455f-9896-620ce22d28ad","size":2147483648,"type":"ROOT","storagePoolType":"NetworkFilesystem","storagePoolUuid":"17b0a8a5-2376-3d11-b60e-31eebeafb217","deviceId":0},"wait":0}}]
>>>> >> >> >> >>> >> >> } to 233845174730255
>>>> >> >> >> >>> >> >> 2013-10-29 11:51:08,296 DEBUG
>>>> >> >> >> >>> >> >> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> >>> >> >> (AgentManager-Handler-15:null) Seq 1-201919186:
>>>> Forwarding
>>>> >> Seq
>>>> >> >> >> >>> >> >> 1-201919186:
>>>> >> >> >> >>> >> >>  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1,
>>>> Flags:
>>>> >> >> >> >>> >> >> 100111,
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >>
>>>> >>
>>>> [{"storage.DestroyCommand":{"vmName":"v-2-VM","volume":{"id":2,"name":"ROOT-2","mountPoint":"/export/primary","path":"9aa8e81d-edb8-4284-a422-a89e9fd4237c","size":2147483648,"type":"ROOT","storagePoolType":"NetworkFilesystem","storagePoolUuid":"17b0a8a5-2376-3d11-b60e-31eebeafb217","deviceId":0},"wait":0}}]
>>>> >> >> >> >>> >> >> } to 233845174730255
>>>> >> >> >> >>> >> >> 2013-10-29 11:51:08,297 DEBUG
>>>> >> >> >> >>> >> >> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> >>> >> >> (AgentManager-Handler-13:null) Seq 2-1168706613:
>>>> Forwarding
>>>> >> >> Seq
>>>> >> >> >> >>> >> >> 2-1168706613:  { Cmd , MgmtId: 233845174730253, via:
>>>> 2,
>>>> >> Ver:
>>>> >> >> v1,
>>>> >> >> >> >>> >> >> Flags:
>>>> >> >> >> >>> >> >> 100111,
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >>
>>>> >>
>>>> [{"storage.DestroyCommand":{"vmName":"r-5-VM","volume":{"id":6,"name":"ROOT-5","mountPoint":"/export/primary","path":"56b215e0-bbb0-455f-9896-620ce22d28ad","size":2147483648,"type":"ROOT","storagePoolType":"NetworkFilesystem","storagePoolUuid":"17b0a8a5-2376-3d11-b60e-31eebeafb217","deviceId":0},"wait":0}}]
>>>> >> >> >> >>> >> >> } to 233845174730255
>>>> >> >> >> >>> >> >> 2013-10-29 11:51:08,297 DEBUG
>>>> >> >> >> >>> >> >> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> >>> >> >> (AgentManager-Handler-10:null) Seq 1-201919186:
>>>> Forwarding
>>>> >> Seq
>>>> >> >> >> >>> >> >> 1-201919186:
>>>> >> >> >> >>> >> >>  { Cmd , MgmtId: 233845174730253, via: 1, Ver: v1,
>>>> Flags:
>>>> >> >> >> >>> >> >> 100111,
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >>
>>>> >>
>>>> [{"storage.DestroyCommand":{"vmName":"v-2-VM","volume":{"id":2,"name":"ROOT-2","mountPoint":"/export/primary","path":"9aa8e81d-edb8-4284-a422-a89e9fd4237c","size":2147483648,"type":"ROOT","storagePoolType":"NetworkFilesystem","storagePoolUuid":"17b0a8a5-2376-3d11-b60e-31eebeafb217","deviceId":0},"wait":0}}]
>>>> >> >> >> >>> >> >> } to 233845174730255
>>>> >> >> >> >>> >> >> 2013-10-29 11:51:08,299 DEBUG
>>>> >> >> >> >>> >> >> [agent.manager.ClusteredAgentAttache]
>>>> >> >> >> >>> >> >> (AgentManager-Handler-3:null) Seq 2-1168706613:
>>>> Forwarding
>>>> >> Seq
>>>> >> >> >> >>> >> >> 2-1168706613:  { Cmd , MgmtId: 233845174730253, via:
>>>> 2,
>>>> >> Ver:
>>>> >> >> v1,
>>>> >> >> >> >>> >> >> Flags:
>>>> >> >> >> >>> >> >> 100111,
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >>
>>>> >>
>>>> [{"storage.DestroyCommand":{"vmName":"r-5-VM","volume":{"id":6,"name":"ROOT-5","mountPoint":"/export/primary","path":"56b215e0-bbb0-455f-9896-620ce22d28ad","size":2147483648,"type":"ROOT","storagePoolType":"NetworkFilesystem","storagePoolUuid":"17b0a8a5-2376-3d11-b60e-31eebeafb217","deviceId":0},"wait":0}}]
>>>> >> >> >> >>> >> >> } to 233845174730255
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >> >>
>>>> >> >> >> >>> >
>>>> >> >> >> >>> >
>>>> >> >> >> >>
>>>> >> >> >> >>
>>>> >> >> >> >
>>>> >> >> >
>>>> >> >> >
>>>> >> >>
>>>> >>
>>>>
>>>
>>>
>>

Reply via email to