RCE found in log4j library

2021-12-10 Thread Rakesh Venkatesh
Hello users/devs


Recently there was an RCE found in log4j library. Fortunately we are
running on older version and so we are good I guess. But if we plan to
upgrade in future, we need to keep the library version in mind

https://www.lunasec.io/docs/blog/log4j-zero-day/

-- 
Thanks and regards
Rakesh


How to use @Configuration annotation

2021-11-24 Thread Rakesh Venkatesh
Hello devs


I am trying to integrate third party library into cloudstack but the
library supports  @Configuration and @Enable* annotations to use it and not
with xml files.

I tried adding that annotation on a config file, built a new package but
the cloudstack wont boot.
It is stuck at below line

2021-11-24 11:40:50,554 INFO  [o.a.c.s.m.m.i.DefaultModuleDefinitionSet]
(main:null) (logid:) Loading module context [core] from URL
[jar:file:/usr/share/cloudstack-management/lib/cloudstack-4.16.1.0-SNAPSHOT.jar!/META-INF/cloudstack/system/spring-contrail-system-context-inheritable.xml]
2021-11-24 11:40:50,554 INFO  [o.a.c.s.m.m.i.DefaultModuleDefinitionSet]
(main:null) (logid:) Loading module context [core] from URL
[jar:file:/usr/share/cloudstack-management/lib/cloudstack-4.16.1.0-SNAPSHOT.jar!/META-INF/cloudstack/bootstrap/spring-bootstrap-context-inheritable.xml]


Is there anything else I need to do to enable this feature? I also added
bean id definition in one of the xml files under server/ but that also
didnt help.

-- 
Thanks and regards
Rakesh


api to import network acl rules

2021-10-11 Thread Rakesh Venkatesh
Hello

I have a network acl which has around 800 rules. i want to create another
network acl with these rules along with some extra rules. Is there a way to
just copy all these 800 rules from one network acl to another one?

I see the export acl's button but there is no option to import it. Is there
any other simple way?

-- 
Thanks and regards
Rakesh


2FA

2021-08-10 Thread Rakesh Venkatesh
Hello

Has anyone thought about 2FA or about how to implement it in cloudstack?
Looks like this will be good addition to enhance the security. I have some
idea about implementing in the backend but dont have much idea on how to
display the QR code in ui or other functionalities which is needed for
frontend part.

-- 
Thanks and regards
Rakesh


Re: Limiting the conntrackd logs in virtual router

2021-07-28 Thread Rakesh Venkatesh
Hello Andrija

Sorry for replying to your email this late. Just saw it.
Basically there was an issue with logrotate service as it was running just
once a day instead of hourly. I have created a pr for that in 4.15

On Tue, Jun 15, 2021 at 12:41 AM Andrija Panic 
wrote:

> Rakesh,
>
> can you confirm if you are seeing this for ALL VPCs with S-2-S enabled
> can you describe/quantify the MB/s log increase over time - does it have
> anything to do with the traffic throughput through the tunnel, or is it
> just happening even with no traffic (i.e. you can setup S-2-S tunnel
> between 2 VPCs (one of them marked as "passive") - I would like to see/test
> this in 4.15.1 (if you confirm that the amount of traffic does NOT
> influence the size of the logs)
>
> Please respond ASAP - as we need to craft 4.15.1 RC2 - if there is a bug,
> we need to fix it (and I will help testing).
>
> Best,
>
>
> On Wed, 9 Jun 2021 at 13:31, Rakesh Venkatesh 
> wrote:
>
> > If I restart with cleanup then the modified config will be deleted and
> the
> > log starts populating very fast again.
> >
> > On Wed, Jun 9, 2021 at 1:19 PM Andrija Panic 
> > wrote:
> >
> > > Did you try restarting that VPC with cleanup - so brand new VR is
> > created -
> > > to see if the issue is still observed.
> > >
> > > Can you test the same now with 4.15.1 RC1 (voting is happening, please
> > feel
> > > free to test and participate!)
> > >
> > > Best,
> > >
> > > On Wed, 9 Jun 2021 at 10:57, Rakesh Venkatesh <
> www.rakeshv@gmail.com
> > >
> > > wrote:
> > >
> > > > I have 4.14 version. The issue exists with only 1 customer.
> > > > The VR's belonging to other customers are fine but for one customer
> who
> > > has
> > > > enabled site 2 site vpn, its creating lot of logs.
> > > > The config I mentioned above is a temporary fix as it will be deleted
> > > once
> > > > the network is restarted with cleanup.
> > > >
> > > > On Wed, Jun 9, 2021 at 10:52 AM Andrija Panic <
> andrija.pa...@gmail.com
> > >
> > > > wrote:
> > > >
> > > > > Can you advise which version and have you tried isolating the
> > problem -
> > > > > this is first time I hear for such an issue (and I have customers
> > with
> > > 3y
> > > > > VR uptime with no issues)
> > > > >
> > > > > Perhaps something is broken inside that VR, thus the excessive
> > logging,
> > > > > etc?
> > > > >
> > > > > Best,
> > > > >
> > > > > On Wed, 9 Jun 2021 at 10:07, Rakesh Venkatesh <
> > > www.rakeshv@gmail.com
> > > > >
> > > > > wrote:
> > > > >
> > > > > > Hello
> > > > > >
> > > > > > I have an issue in the virtual router where
> > > conntrackd(conntrack-tools)
> > > > > is
> > > > > > populating /var/log/daemon.log a lot and because of that VR is
> > > running
> > > > > out
> > > > > > of disk space. Do you guys know how to prevent conntrackd from
> > > logging
> > > > > too
> > > > > > much? I manually changed the logrotate.conf to rotate every hour
> > and
> > > > > retain
> > > > > > only few copies but thats a quick fix. Is there any better fix to
> > > avoid
> > > > > > this issue? In our setup the VR has 2GB disk and even that got
> > filled
> > > > up
> > > > > > pretty quickly.
> > > > > >
> > > > > > The logrotate config for daemon.log is present in
> > > > > /etc/logrotate.d/rsyslog
> > > > > > and I added
> > > > > >
> > > > > > /var/log/daemon.log
> > > > > > {
> > > > > > hourly
> > > > > > rotate 1
> > > > > > missingok
> > > > > > notifempty
> > > > > > compress
> > > > > > copytruncate
> > > > > > }
> > > > > >
> > > > > >
> > > > > > The config present in /etc/logrotate.d/conntrackd is pointing
> > > > > > to /var/log/conntrackd-stats.log and so changing that config wont
> > > help.
> > > > > >
> > > > > > --
> > > > > > Thanks and regards
> > > > > > Rakesh
> > > > > >
> > > > >
> > > > >
> > > > > --
> > > > >
> > > > > Andrija Panić
> > > > >
> > > >
> > > >
> > > > --
> > > > Thanks and regards
> > > > Rakesh venkatesh
> > > >
> > >
> > >
> > > --
> > >
> > > Andrija Panić
> > >
> >
> >
> > --
> > Thanks and regards
> > Rakesh venkatesh
> >
>
>
> --
>
> Andrija Panić
>


-- 
Thanks and regards
Rakesh venkatesh


Code executed twice while deploying VM

2021-07-28 Thread Rakesh Venkatesh
Hello Users/Devs


Today I observed that while deploying a VM, the same code is executed
twice: the first time, while trying to find a suitable deployment
destination and the second time, while trying to start a VM. I think this
is a redundant process and also time-consuming since the same calculations
are done twice. These are the one of the duplicate logs

Asyn job is created
2021-07-28 09:01:27,964 (logid:10ac9f28) ZoneWideStoragePoolAllocator to
find storage pool

Now it finds a suitable destination
Returning Deployment Destination:
Dest[Zone(Id)-Pod(Id)-Cluster(Id)-Host(Id)-Storage(Volume(Id|Type-->Pool(Id))]
: Dest[Zone(1)-Pod(1)-Cluster(1)-Host(1)-Storage(Volume(23|ROOT-->Pool(3))]

VM start attempt #1

2021-07-28 09:01:36,039 job-174/job-175 ctx-b815930d) (logid:10ac9f28)
ZoneWideStoragePoolAllocator to find storage pool


>From the above two logs, I can see that starting at 9:01:27 and 09:01:36,
its doing the same calculations.
if I see the logs before and after the above-mentioned logs, they are all
same. They try to find a suitable host, a suitable storage pool, and so on.
In the production platform, the calculation is done twice and this takes a
lot of time to deploy a VM.

Do you guys also think this is an issue or redundant process and needs to
be improved? Or any other suggestions to avoid double calculation?

-- 
Thanks and regards
Rakesh


Disabling a storage pool

2021-06-29 Thread Rakesh Venkatesh
Hello folks

Is there a way to disable a particular storage pool so that it won't be
used for further volume allocation? I don't want to enable the maintenance
mode as that will turn off the VM's whose volumes running on that pool. I
don't want to use a global setting also since this will come into effect
after the threshold value is reached.

In some cases even if the pool is just 10% allocated, I still want to
disable it so that the current volumes will keep existing on the same pool
and at the same time further deployment of volumes on this pool is disabled.

I looked at the storge tags options but that involves adding tags to
service offerings and I dont want to mess up with that tags. Should we add
a new api to enable this feature? or any other better suggestion?

-- 
Thanks and regards
Rakesh


Re: Keyboard Shortcuts on Management Server.

2021-06-09 Thread Rakesh Venkatesh
I looked at the GSOC idea. I have feedback on this

1. Need to memorize the shortcut which is not so good
2. only a limited set of shortcuts as of now.. if it grows then more
memorization


If we use vimium or similar plugin


cons

need to spend time installing in browsers


pros

very less memorization
more shortcuts
no need for source code change if more tabs/panes or whatever is added to
ui future. (just press F and you are good to go)

On Wed, Jun 9, 2021 at 1:38 PM Boris Stoyanov 
wrote:

> Not everyone uses vimium browser plugin, I think it would be great if we
> can get this working cross-browser.
>
> Here’s the original GSOC Idea:
> https://github.com/apache/cloudstack/issues/4798
>
> Bobby.
>
> From: Rakesh Venkatesh 
> Date: Wednesday, 9 June 2021, 14:29
> To: dev 
> Subject: Re: Keyboard Shortcuts on Management Server.
> I usually use vimium plugin which has keyboard shortcuts for all links in a
> webpage which is much more handy
>
>
>
>
> On Wed, Jun 9, 2021 at 12:38 PM Apurv Gupta  wrote:
>
> > Hi all,
> >
> >
> >
> > I’m trying to implement the keyboard shortcuts and the idea is to use the
> > management server ui without a mouse.
> >
> > From the mentors, here are the suggestion list for shortcuts:
> >
> >
> >
> > [i] - infrastructure
> >
> > [g] - configuration->global settings
> >
> > [v] - storage->volumes
> >
> > [c] - compute->instances
> >
> > [t] - images->templates
> >
> > [n] - networks
> >
> > [a] - accounts
> >
> > [p] - projects
> >
> > [e] - events
> >
> >
> >
> > Please feel free to add more to the list and would appreciate any
> comments,
> > suggestions or objections.
> > Also, I’ve opened a draft PR #5090 (
> > https://github.com/apache/cloudstack/pull/5090).
> >
> >
> >
> > Thankyou!
> >
>
>
> --
> Thanks and regards
> Rakesh venkatesh
>


-- 
Thanks and regards
Rakesh venkatesh


Re: Limiting the conntrackd logs in virtual router

2021-06-09 Thread Rakesh Venkatesh
If I restart with cleanup then the modified config will be deleted and the
log starts populating very fast again.

On Wed, Jun 9, 2021 at 1:19 PM Andrija Panic 
wrote:

> Did you try restarting that VPC with cleanup - so brand new VR is created -
> to see if the issue is still observed.
>
> Can you test the same now with 4.15.1 RC1 (voting is happening, please feel
> free to test and participate!)
>
> Best,
>
> On Wed, 9 Jun 2021 at 10:57, Rakesh Venkatesh 
> wrote:
>
> > I have 4.14 version. The issue exists with only 1 customer.
> > The VR's belonging to other customers are fine but for one customer who
> has
> > enabled site 2 site vpn, its creating lot of logs.
> > The config I mentioned above is a temporary fix as it will be deleted
> once
> > the network is restarted with cleanup.
> >
> > On Wed, Jun 9, 2021 at 10:52 AM Andrija Panic 
> > wrote:
> >
> > > Can you advise which version and have you tried isolating the problem -
> > > this is first time I hear for such an issue (and I have customers with
> 3y
> > > VR uptime with no issues)
> > >
> > > Perhaps something is broken inside that VR, thus the excessive logging,
> > > etc?
> > >
> > > Best,
> > >
> > > On Wed, 9 Jun 2021 at 10:07, Rakesh Venkatesh <
> www.rakeshv@gmail.com
> > >
> > > wrote:
> > >
> > > > Hello
> > > >
> > > > I have an issue in the virtual router where
> conntrackd(conntrack-tools)
> > > is
> > > > populating /var/log/daemon.log a lot and because of that VR is
> running
> > > out
> > > > of disk space. Do you guys know how to prevent conntrackd from
> logging
> > > too
> > > > much? I manually changed the logrotate.conf to rotate every hour and
> > > retain
> > > > only few copies but thats a quick fix. Is there any better fix to
> avoid
> > > > this issue? In our setup the VR has 2GB disk and even that got filled
> > up
> > > > pretty quickly.
> > > >
> > > > The logrotate config for daemon.log is present in
> > > /etc/logrotate.d/rsyslog
> > > > and I added
> > > >
> > > > /var/log/daemon.log
> > > > {
> > > > hourly
> > > > rotate 1
> > > > missingok
> > > > notifempty
> > > > compress
> > > > copytruncate
> > > > }
> > > >
> > > >
> > > > The config present in /etc/logrotate.d/conntrackd is pointing
> > > > to /var/log/conntrackd-stats.log and so changing that config wont
> help.
> > > >
> > > > --
> > > > Thanks and regards
> > > > Rakesh
> > > >
> > >
> > >
> > > --
> > >
> > > Andrija Panić
> > >
> >
> >
> > --
> > Thanks and regards
> > Rakesh venkatesh
> >
>
>
> --
>
> Andrija Panić
>


-- 
Thanks and regards
Rakesh venkatesh


Re: Limiting the conntrackd logs in virtual router

2021-06-09 Thread Rakesh Venkatesh
I have 4.14 version. The issue exists with only 1 customer.
The VR's belonging to other customers are fine but for one customer who has
enabled site 2 site vpn, its creating lot of logs.
The config I mentioned above is a temporary fix as it will be deleted once
the network is restarted with cleanup.

On Wed, Jun 9, 2021 at 10:52 AM Andrija Panic 
wrote:

> Can you advise which version and have you tried isolating the problem -
> this is first time I hear for such an issue (and I have customers with 3y
> VR uptime with no issues)
>
> Perhaps something is broken inside that VR, thus the excessive logging,
> etc?
>
> Best,
>
> On Wed, 9 Jun 2021 at 10:07, Rakesh Venkatesh 
> wrote:
>
> > Hello
> >
> > I have an issue in the virtual router where conntrackd(conntrack-tools)
> is
> > populating /var/log/daemon.log a lot and because of that VR is running
> out
> > of disk space. Do you guys know how to prevent conntrackd from logging
> too
> > much? I manually changed the logrotate.conf to rotate every hour and
> retain
> > only few copies but thats a quick fix. Is there any better fix to avoid
> > this issue? In our setup the VR has 2GB disk and even that got filled up
> > pretty quickly.
> >
> > The logrotate config for daemon.log is present in
> /etc/logrotate.d/rsyslog
> > and I added
> >
> > /var/log/daemon.log
> > {
> > hourly
> > rotate 1
> > missingok
> > notifempty
> > compress
> > copytruncate
> > }
> >
> >
> > The config present in /etc/logrotate.d/conntrackd is pointing
> > to /var/log/conntrackd-stats.log and so changing that config wont help.
> >
> > --
> > Thanks and regards
> > Rakesh
> >
>
>
> --
>
> Andrija Panić
>


-- 
Thanks and regards
Rakesh venkatesh


Limiting the conntrackd logs in virtual router

2021-06-09 Thread Rakesh Venkatesh
Hello

I have an issue in the virtual router where conntrackd(conntrack-tools) is
populating /var/log/daemon.log a lot and because of that VR is running out
of disk space. Do you guys know how to prevent conntrackd from logging too
much? I manually changed the logrotate.conf to rotate every hour and retain
only few copies but thats a quick fix. Is there any better fix to avoid
this issue? In our setup the VR has 2GB disk and even that got filled up
pretty quickly.

The logrotate config for daemon.log is present in /etc/logrotate.d/rsyslog
and I added

/var/log/daemon.log
{
hourly
rotate 1
missingok
notifempty
compress
copytruncate
}


The config present in /etc/logrotate.d/conntrackd is pointing
to /var/log/conntrackd-stats.log and so changing that config wont help.

-- 
Thanks and regards
Rakesh


SSL handshake error while connecting to mgt server

2021-05-20 Thread Rakesh Venkatesh
Hello

Host is failing to connect to mgt server with the following error message.
Any idea what could be the reason. It started happening all of a sudden. I
have other hosts in the platform and they are working fine.

2021-05-20 12:34:31,725 ERROR [utils.nio.NioConnection] (main:null) Unable
to initialize the threads.
javax.net.ssl.SSLHandshakeException: No appropriate protocol (protocol is
disabled or cipher suites are inappropriate)

> at sun.security.ssl.HandshakeContext.(HandshakeContext.java:171)
>
> at
> sun.security.ssl.ClientHandshakeContext.(ClientHandshakeContext.java:98)
> at sun.security.ssl.TransportContext.kickstart(TransportContext.java:220)
> at sun.security.ssl.SSLEngineImpl.writeRecord(SSLEngineImpl.java:159)
> at sun.security.ssl.SSLEngineImpl.wrap(SSLEngineImpl.java:130)
> at sun.security.ssl.SSLEngineImpl.wrap(SSLEngineImpl.java:110)
> at javax.net.ssl.SSLEngine.wrap(SSLEngine.java:471)
> at com.cloud.utils.nio.Link.doHandshake(Link.java:489)
> at com.cloud.utils.nio.NioClient.init(NioClient.java:80)
> at com.cloud.utils.nio.NioConnection.start(NioConnection.java:88)
> at com.cloud.agent.Agent.start(Agent.java:246)
> at com.cloud.agent.AgentShell.launchAgent(AgentShell.java:448)
> at com.cloud.agent.AgentShell.launchAgentFromClassInfo(AgentShell.java:416)
> at com.cloud.agent.AgentShell.launchAgent(AgentShell.java:400)
> at com.cloud.agent.AgentShell.start(AgentShell.java:510)
> at com.cloud.agent.AgentShell.main(AgentShell.java:545)
>
>
-- 
Thanks and regards
Rakesh


Changing the role of the account

2021-04-22 Thread Rakesh Venkatesh
Hello folks

I dont think there is an endpoint right now to change the role of the
account. If I have to change that in db, that should be doable right by
changing the id's in two or three tables. I hope that wont break anything
else. Is it good to add a new parameter to updateAccount api to take the
new role id?

-- 
Thanks and regards
Rakesh venkatesh


Re: Db schema upgrade fails from 4.14 to 4.16

2021-04-09 Thread Rakesh Venkatesh
Hello Pearl


No, this was the first time I tried to upgrade from 4.14 to master. 4.14
was running for quite long time.

I took db backup, registered 4.16 systemvm, changed global settings and
then just deployed new packages. It failed in the first attempt.
I was able to resolve this by dropping the project_account and
project_invitation table, creating them and restarted the mgt service. This
time it applied successfully


To double check it, I created a fresh 4.14 system, did the above steps and
the upgrade still failed.
So solution right now is to drop the tables, create it again and run the
upgrade

On Thu, Apr 8, 2021 at 6:52 PM Pearl d'Silva 
wrote:

> Hi Rakesh,
>
> Was there a prior attempt made to upgrade which didn't go through
> successfully? It seems like while it failed to upgrade, the DB changes were
> applied, hence resulting in the failure that is noticed. To address the
> issue, you will have to restore the DB, if you've taken a backup of it and
> redo the upgrade.
> However, to avoid such situations, an enhancement was added with
> https://github.com/apache/cloudstack/pull/4582 which prevents upgrading
> the DB in case of missing systemvm template.
>
> Thanks
> Pearl
>
> 
> From: Andrija Panic 
> Sent: Thursday, April 8, 2021 9:09 PM
> To: dev 
> Cc: users 
> Subject: Re: Db schema upgrade fails from 4.14 to 4.16
>
> 4.16 doesn't exist yet, and it's the master branch - so failures are
> expected...
>
> On Thu, 8 Apr 2021 at 16:42, Rakesh Venkatesh  http://www.rakeshv@gmail.com>>
> wrote:
>
> > Hello users
> >
> >
> > Im trying to upgrade cloudstack from 4.14.0 to 4.16.0 and the upgrade is
> > failing because of the db schema upgrade. Any idea on how to fix this.
> >
> > This is the commit which is causing the issue
> > https://github.com/apache/cloudstack/pull/4128
> >
> >
> >
> > 2021-04-08 13:59:26,983 DEBUG [c.c.u.d.ScriptRunner] (main:null) (logid:)
> > -- Schema upgrade from 4.14.0.0 to 4.15.0.0
> >
> > 2021-04-08 13:59:27,002 ERROR [c.c.u.d.ScriptRunner] (main:null) (logid:)
> > Error executing: ALTER TABLE `cloud`.`project_account`  ADD COLUMN
> > `user_id` bigint unsigned COMMENT 'ID of user to be added to the project'
> > AFTER `account_id`,  ADD CONSTRAINT `fk_project_account__user_id` FOREIGN
> > KEY `fk_project_account__user_id`(`user_id`) REFERENCES `user`(`id`) ON
> > DELETE CASCADE,  ADD COLUMN `project_role_id` bigint unsigned COMMENT
> > 'Project role id' AFTER `project_account_id`,  ADD CONSTRAINT
> > `fk_project_account__project_role_id` FOREIGN KEY (`project_role_id`)
> > REFERENCES `project_role` (`id`) ON DELETE SET NULL,  DROP FOREIGN KEY
> > `fk_project_account__account_id`,  DROP INDEX `account_id`
> > 2021-04-08 13:59:27,003 ERROR [c.c.u.d.ScriptRunner] (main:null) (logid:)
> > java.sql.SQLSyntaxErrorException: Can't DROP
> > 'fk_project_account__account_id'; check that column/key exists
> > 2021-04-08 13:59:27,004 ERROR [c.c.u.DatabaseUpgradeChecker] (main:null)
> > (logid:) Unable to execute upgrade script
> > java.sql.SQLSyntaxErrorException: Can't DROP
> > 'fk_project_account__account_id'; check that column/key exists
> > at com.cloud.utils.db.ScriptRunner.runScript(ScriptRunner.java:185)
> >
> >
> > Caused by: java.sql.SQLSyntaxErrorException: Can't DROP
> > 'fk_project_account__account_id'; check that column/key exists
> > --
> > Thanks and regards
> > Rakesh
> >
>
>
> --
>
> Andrija Panić
>
> pearl.dsi...@shapeblue.com
> www.shapeblue.com
> 3 London Bridge Street,  3rd floor, News Building, London  SE1 9SGUK
> @shapeblue
>
>
>
>

-- 
Thanks and regards
Rakesh venkatesh


Db schema upgrade fails from 4.14 to 4.16

2021-04-08 Thread Rakesh Venkatesh
Hello users


Im trying to upgrade cloudstack from 4.14.0 to 4.16.0 and the upgrade is
failing because of the db schema upgrade. Any idea on how to fix this.

This is the commit which is causing the issue
https://github.com/apache/cloudstack/pull/4128



2021-04-08 13:59:26,983 DEBUG [c.c.u.d.ScriptRunner] (main:null) (logid:)
-- Schema upgrade from 4.14.0.0 to 4.15.0.0

2021-04-08 13:59:27,002 ERROR [c.c.u.d.ScriptRunner] (main:null) (logid:)
Error executing: ALTER TABLE `cloud`.`project_account`  ADD COLUMN
`user_id` bigint unsigned COMMENT 'ID of user to be added to the project'
AFTER `account_id`,  ADD CONSTRAINT `fk_project_account__user_id` FOREIGN
KEY `fk_project_account__user_id`(`user_id`) REFERENCES `user`(`id`) ON
DELETE CASCADE,  ADD COLUMN `project_role_id` bigint unsigned COMMENT
'Project role id' AFTER `project_account_id`,  ADD CONSTRAINT
`fk_project_account__project_role_id` FOREIGN KEY (`project_role_id`)
REFERENCES `project_role` (`id`) ON DELETE SET NULL,  DROP FOREIGN KEY
`fk_project_account__account_id`,  DROP INDEX `account_id`
2021-04-08 13:59:27,003 ERROR [c.c.u.d.ScriptRunner] (main:null) (logid:)
java.sql.SQLSyntaxErrorException: Can't DROP
'fk_project_account__account_id'; check that column/key exists
2021-04-08 13:59:27,004 ERROR [c.c.u.DatabaseUpgradeChecker] (main:null)
(logid:) Unable to execute upgrade script
java.sql.SQLSyntaxErrorException: Can't DROP
'fk_project_account__account_id'; check that column/key exists
at com.cloud.utils.db.ScriptRunner.runScript(ScriptRunner.java:185)


Caused by: java.sql.SQLSyntaxErrorException: Can't DROP
'fk_project_account__account_id'; check that column/key exists
-- 
Thanks and regards
Rakesh


Automatically applying DB schema changes

2020-10-16 Thread Rakesh Venkatesh
Hello Users and Dev

Is there a way a new DB schema changes can be applied automatically
whenever I install new packages? My setup was running with two month old
changes of 4.15 and when I deployed new packages with latest changes, all
the recent db scheme changes are not applied and I need to run it manually.
How do I avoid it and how do you guys do it?

For example: This is the error I get

Caused by: java.sql.SQLSyntaxErrorException: Unknown column
'image_store.readonly' in 'field list'

and a big stack trace

This was fixed by applying changes from
https://github.com/apache/cloudstack/blob/master/engine/schema/src/main/resources/META-INF/db/schema-41400to41500.sql#L198-L222


Another error was

Caused by: java.sql.SQLSyntaxErrorException: Unknown column
'project_invitations.account_role' in 'field list'


So I had to apply the schema needed for project_role related queries
-- 
Thanks and regards
Rakesh


Deploying Linux/Windows Vm's on hypervisors

2020-09-15 Thread Rakesh Venkatesh
Hello All


I want your opinion on how you organize deploying different types of
VM's(Linux, windows) on a specific set of hypervisors.

In our setup, we have added a tag "Linux,Unix" on few hypervisors and also
on templates so that when we create a vm using these templates then it will
be deployed on HV's matching that tag only. If there is no tag on the
template then it can be deployed on any hypervisor.

We have some windows templates which do not have a tag as they are uploaded
by customers and when VM is created from that template, it will be deployed
in any suitable hypervisor but we want it to be deployed on a hypervisor
having "Windows" tag. Is there any way to achieve this apart from using
template tag?

Another way is to use the OS type and match with hypervisor tags but that
may not be a suitable solution. So if anyone has managed to deploy windows
vm's on windows tagged hypervisor and linux vm's on linux/unix tagged
hypervisor then please let me know
-- 
Thanks and regards
Rakesh


Re: Unable to ping secondary NIC ip of a VM in VPC

2020-08-13 Thread Rakesh Venkatesh
Hi Andrija

Yes if I configure eth1 manually inside VM to get Ip from dhcp then ping
works fine but not sure how to do this inside windows vm

On Wed, Aug 12, 2020 at 10:43 PM Andrija Panic 
wrote:

> Hi Rakesh,
>
> if you configure eth1 manually inside a VM, but set it to obtain the IP via
> DHCP, does it obtain the IP properly then?
>
> Best,
>
> On Tue, 11 Aug 2020, 12:41 Rakesh Venkatesh, 
> wrote:
>
> > Wei,
> >
> > This was the test done on ubuntu os. The initial issue was found on
> windows
> > vm.
> >
> > Even if it can be solved by adding some config inside the vm, do we need
> to
> > do every time we delete/add additional NIC's to vm? That will be extra
> work
> > all the time right?
> >
> > On Mon, Aug 10, 2020 at 5:43 PM Wei ZHOU  wrote:
> >
> > > Hi Rakesh,
> > >
> > > It is not a bug. you need to configure the networks inside vm. for
> > example
> > > /etc/network/interfaces in old ubuntu versions or /etc/netplan in newer
> > > ubuntu versions.
> > >
> > > -Wei
> > >
> > > On Monday, 10 August 2020, Rakesh Venkatesh  >
> > > wrote:
> > >
> > > > Hello Thomas
> > > >
> > > >
> > > > Thanks for the reply. It looks like the NIC is not configured
> properly
> > in
> > > > VM. I dont see IP 10.10.2.23 being configured
> > > >
> > > > # ip a
> > > > 1: lo:  mtu 65536 qdisc noqueue state UNKNOWN
> > group
> > > > default qlen 1
> > > > link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> > > > inet 127.0.0.1/8 scope host lo
> > > >valid_lft forever preferred_lft forever
> > > > inet6 ::1/128 scope host
> > > >valid_lft forever preferred_lft forever
> > > > 2: eth0:  mtu 1500 qdisc pfifo_fast
> > > state
> > > > UP group default qlen 1000
> > > > link/ether 02:00:78:24:00:01 brd ff:ff:ff:ff:ff:ff
> > > > inet 10.10.1.51/24 brd 10.10.1.255 scope global eth0
> > > >valid_lft forever preferred_lft forever
> > > > inet6 fe80::78ff:fe24:1/64 scope link
> > > >valid_lft forever preferred_lft forever
> > > > 3: eth1:  mtu 1500 qdisc noop state DOWN group
> > > default
> > > > qlen 1000
> > > > link/ether 02:00:1f:ed:00:0e brd ff:ff:ff:ff:ff:ff
> > > >
> > > >
> > > >
> > > > I removed the secondary NIC, stopped the vm, added secondary NIC and
> > > > started the VM. The extra NIC is still not configured in VM.
> > > > Is this a bug?
> > > >
> > > > On Mon, Aug 10, 2020 at 4:46 PM Thomas Joseph <
> thomas.jo...@gmail.com>
> > > > wrote:
> > > >
> > > > > Hello Venkatesh,
> > > > >
> > > > > I'm guessing from the output provided that you have tried to ping
> the
> > > ip
> > > > > from the VR.
> > > > > Have you checked if the NIC is up on the VM?
> > > > >  you can run tcpdump on the related VM to check if the icmp packets
> > are
> > > > > reaching the VM.
> > > > >
> > > > > Regards
> > > > > Thomas
> > > > >
> > > > > On Mon, 10 Aug 2020, 8:06 pm Rakesh Venkatesh, <
> > > > www.rakeshv@gmail.com>
> > > > > wrote:
> > > > >
> > > > > > Hello,
> > > > > >
> > > > > > I have created a VPC with three tiers in it (tier1, tier2. tier3)
> > > using
> > > > > > cloudstack 4.14.0.0 version with Ubuntu16 as OS for mgt, agent
> > > > > > I created a VM in tier1 and I am able to ping its IP address
> > > > > >
> > > > > > # ping 10.10.1.51
> > > > > > PING 10.10.1.51 (10.10.1.51): 56 data bytes
> > > > > > 64 bytes from 10.10.1.51: icmp_seq=0 ttl=64 time=2.273 ms
> > > > > > 64 bytes from 10.10.1.51: icmp_seq=1 ttl=64 time=1.690 ms
> > > > > > ^C--- 10.10.1.51 ping statistics ---
> > > > > >
> > > > > >
> > > > > > After adding secondary NIC to vm from tier2 or tier3 then I am
> not
> > > able
> > > > > to
> > > > > > ping that IP address.
> > > > > >
> > > > > > # ping 10.10.2.23
> > > &g

Re: How to delete OS template via database

2020-08-12 Thread Rakesh Venkatesh
vm_template is not an api. Its a table in cloud database

On Wed, Aug 12, 2020 at 11:35 AM Wan Rahim, Wan Mohd Saifudin <
wanmohdsaifudin.wanra...@itelligence.com.my> wrote:

> Already tested to delete the template in cloudmonkey but still failed.
>
> And vm_template api is not available.
>
> Is there any other method can be used?
>
> Thanks
> Din
>
> -Original Message-
> From: Rakesh Venkatesh [mailto:www.rakeshv@gmail.com]
> Sent: Tuesday, 11 August, 2020 6:47 PM
> To: users 
> Subject: Re: How to delete OS template via database
>
> EXTERNAL MAIL:
>
> You can set "state" to "Inactive" in vm_template table and it wont show up
> in UI
>
> update vm_template set state="Inactive" where id=<> limit 1;
>
> or you can run delete template id= cloudmonkey api to delete it
>
> On Tue, Aug 11, 2020 at 9:49 AM harith roli  wrote:
>
> > please help, I need a way to delete OS template of cloudstack via
> > database. Currently the status of template is UploadAbandoned. I try
> > delete using gui, Zones --> click button "x" to delete template  but I
> > got message box indicate status failed to delete template. Thank you
> > for your help
> >
>
>
> --
> Thanks and regards
> Rakesh venkatesh
>
>
>


-- 
Thanks and regards
Rakesh venkatesh


Re: How to delete OS template via database

2020-08-11 Thread Rakesh Venkatesh
You can set "state" to "Inactive" in vm_template table and it wont show up
in UI

update vm_template set state="Inactive" where id=<> limit 1;

or you can run delete template id= cloudmonkey api to delete it

On Tue, Aug 11, 2020 at 9:49 AM harith roli  wrote:

> please help, I need a way to delete OS template of cloudstack via
> database. Currently the status of template is UploadAbandoned. I try delete
> using gui, Zones --> click button "x" to delete template  but I got message
> box indicate status failed to delete template. Thank you for your help
>


-- 
Thanks and regards
Rakesh venkatesh


Re: Unable to ping secondary NIC ip of a VM in VPC

2020-08-11 Thread Rakesh Venkatesh
Wei,

This was the test done on ubuntu os. The initial issue was found on windows
vm.

Even if it can be solved by adding some config inside the vm, do we need to
do every time we delete/add additional NIC's to vm? That will be extra work
all the time right?

On Mon, Aug 10, 2020 at 5:43 PM Wei ZHOU  wrote:

> Hi Rakesh,
>
> It is not a bug. you need to configure the networks inside vm. for example
> /etc/network/interfaces in old ubuntu versions or /etc/netplan in newer
> ubuntu versions.
>
> -Wei
>
> On Monday, 10 August 2020, Rakesh Venkatesh 
> wrote:
>
> > Hello Thomas
> >
> >
> > Thanks for the reply. It looks like the NIC is not configured properly in
> > VM. I dont see IP 10.10.2.23 being configured
> >
> > # ip a
> > 1: lo:  mtu 65536 qdisc noqueue state UNKNOWN group
> > default qlen 1
> > link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> > inet 127.0.0.1/8 scope host lo
> >valid_lft forever preferred_lft forever
> > inet6 ::1/128 scope host
> >valid_lft forever preferred_lft forever
> > 2: eth0:  mtu 1500 qdisc pfifo_fast
> state
> > UP group default qlen 1000
> > link/ether 02:00:78:24:00:01 brd ff:ff:ff:ff:ff:ff
> > inet 10.10.1.51/24 brd 10.10.1.255 scope global eth0
> >valid_lft forever preferred_lft forever
> > inet6 fe80::78ff:fe24:1/64 scope link
> >valid_lft forever preferred_lft forever
> > 3: eth1:  mtu 1500 qdisc noop state DOWN group
> default
> > qlen 1000
> > link/ether 02:00:1f:ed:00:0e brd ff:ff:ff:ff:ff:ff
> >
> >
> >
> > I removed the secondary NIC, stopped the vm, added secondary NIC and
> > started the VM. The extra NIC is still not configured in VM.
> > Is this a bug?
> >
> > On Mon, Aug 10, 2020 at 4:46 PM Thomas Joseph 
> > wrote:
> >
> > > Hello Venkatesh,
> > >
> > > I'm guessing from the output provided that you have tried to ping the
> ip
> > > from the VR.
> > > Have you checked if the NIC is up on the VM?
> > >  you can run tcpdump on the related VM to check if the icmp packets are
> > > reaching the VM.
> > >
> > > Regards
> > > Thomas
> > >
> > > On Mon, 10 Aug 2020, 8:06 pm Rakesh Venkatesh, <
> > www.rakeshv@gmail.com>
> > > wrote:
> > >
> > > > Hello,
> > > >
> > > > I have created a VPC with three tiers in it (tier1, tier2. tier3)
> using
> > > > cloudstack 4.14.0.0 version with Ubuntu16 as OS for mgt, agent
> > > > I created a VM in tier1 and I am able to ping its IP address
> > > >
> > > > # ping 10.10.1.51
> > > > PING 10.10.1.51 (10.10.1.51): 56 data bytes
> > > > 64 bytes from 10.10.1.51: icmp_seq=0 ttl=64 time=2.273 ms
> > > > 64 bytes from 10.10.1.51: icmp_seq=1 ttl=64 time=1.690 ms
> > > > ^C--- 10.10.1.51 ping statistics ---
> > > >
> > > >
> > > > After adding secondary NIC to vm from tier2 or tier3 then I am not
> able
> > > to
> > > > ping that IP address.
> > > >
> > > > # ping 10.10.2.23
> > > > PING 10.10.2.23 (10.10.2.23): 56 data bytes
> > > > 92 bytes from 10.10.2.254: Destination Host Unreachable
> > > > 92 bytes from 10.10.2.254: Destination Host Unreachable
> > > > 92 bytes from 10.10.2.254: Destination Host Unreachable
> > > > 92 bytes from 10.10.2.254: Destination Host Unreachable
> > > > ^C--- 10.10.2.23 ping statistics ---
> > > >
> > > >
> > > >
> > > > Dchphost content
> > > >
> > > > # cat /etc/dhcphosts.txt
> > > > 02:00:72:36:00:01,10.10.2.83,tier12,infinite
> > > > 02:00:2f:16:00:0c,set:10_10_2_23,10.10.2.23,tier11,infinite
> > > > 02:00:78:24:00:01,10.10.1.51,tier11,infinite
> > > >
> > > >
> > > >
> > > > Dhcpopts content
> > > >
> > > >
> > > > # cat /etc/dhcpopts.txt
> > > > 10_10_2_23,3
> > > > 10_10_2_23,6
> > > > 10_10_2_23,15
> > > >
> > > >
> > > > I don't see any errors in /var/log/cloud.log on VR when secondary nic
> > is
> > > > added.
> > > > Is there any other I can troubleshoot this or look for logs?
> > > >
> > > > --
> > > > Thanks and regards
> > > > Rakesh
> > > >
> > >
> >
> >
> > --
> > Thanks and regards
> > Rakesh venkatesh
> >
>


-- 
Thanks and regards
Rakesh venkatesh


Re: Unable to ping secondary NIC ip of a VM in VPC

2020-08-10 Thread Rakesh Venkatesh
Ok no problem.

As suggested, I manually edited /etc/network/interfaces files to add eth1
and restarted networking service.
Ping works fine now from VR


# ping 10.10.2.25
PING 10.10.2.25 (10.10.2.25): 56 data bytes
64 bytes from 10.10.2.25: icmp_seq=0 ttl=64 time=1.981 ms
64 bytes from 10.10.2.25: icmp_seq=1 ttl=64 time=1.322 ms
64 bytes from 10.10.2.25: icmp_seq=2 ttl=64 time=1.124 ms
64 bytes from 10.10.2.25: icmp_seq=3 ttl=64 time=1.376 ms
^C--- 10.10.2.25 ping statistics ---
4 packets transmitted, 4 packets received, 0% packet loss

On Mon, Aug 10, 2020 at 5:04 PM Thomas Joseph 
wrote:

> Sorry, I've not used v4.14, so someone else can comment on it being a bug.
>
> I would advise you to enable the related nic in the file named interfaces
> and restart the network service to check if that works.
>
>
> On Mon, 10 Aug 2020, 8:29 pm Rakesh Venkatesh, 
> wrote:
>
> > Hello Thomas
> >
> >
> > Thanks for the reply. It looks like the NIC is not configured properly in
> > VM. I dont see IP 10.10.2.23 being configured
> >
> > # ip a
> > 1: lo:  mtu 65536 qdisc noqueue state UNKNOWN group
> > default qlen 1
> > link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> > inet 127.0.0.1/8 scope host lo
> >valid_lft forever preferred_lft forever
> > inet6 ::1/128 scope host
> >valid_lft forever preferred_lft forever
> > 2: eth0:  mtu 1500 qdisc pfifo_fast
> state
> > UP group default qlen 1000
> > link/ether 02:00:78:24:00:01 brd ff:ff:ff:ff:ff:ff
> > inet 10.10.1.51/24 brd 10.10.1.255 scope global eth0
> >valid_lft forever preferred_lft forever
> > inet6 fe80::78ff:fe24:1/64 scope link
> >valid_lft forever preferred_lft forever
> > 3: eth1:  mtu 1500 qdisc noop state DOWN group
> default
> > qlen 1000
> > link/ether 02:00:1f:ed:00:0e brd ff:ff:ff:ff:ff:ff
> >
> >
> >
> > I removed the secondary NIC, stopped the vm, added secondary NIC and
> > started the VM. The extra NIC is still not configured in VM.
> > Is this a bug?
> >
> > On Mon, Aug 10, 2020 at 4:46 PM Thomas Joseph 
> > wrote:
> >
> > > Hello Venkatesh,
> > >
> > > I'm guessing from the output provided that you have tried to ping the
> ip
> > > from the VR.
> > > Have you checked if the NIC is up on the VM?
> > >  you can run tcpdump on the related VM to check if the icmp packets are
> > > reaching the VM.
> > >
> > > Regards
> > > Thomas
> > >
> > > On Mon, 10 Aug 2020, 8:06 pm Rakesh Venkatesh, <
> > www.rakeshv@gmail.com>
> > > wrote:
> > >
> > > > Hello,
> > > >
> > > > I have created a VPC with three tiers in it (tier1, tier2. tier3)
> using
> > > > cloudstack 4.14.0.0 version with Ubuntu16 as OS for mgt, agent
> > > > I created a VM in tier1 and I am able to ping its IP address
> > > >
> > > > # ping 10.10.1.51
> > > > PING 10.10.1.51 (10.10.1.51): 56 data bytes
> > > > 64 bytes from 10.10.1.51: icmp_seq=0 ttl=64 time=2.273 ms
> > > > 64 bytes from 10.10.1.51: icmp_seq=1 ttl=64 time=1.690 ms
> > > > ^C--- 10.10.1.51 ping statistics ---
> > > >
> > > >
> > > > After adding secondary NIC to vm from tier2 or tier3 then I am not
> able
> > > to
> > > > ping that IP address.
> > > >
> > > > # ping 10.10.2.23
> > > > PING 10.10.2.23 (10.10.2.23): 56 data bytes
> > > > 92 bytes from 10.10.2.254: Destination Host Unreachable
> > > > 92 bytes from 10.10.2.254: Destination Host Unreachable
> > > > 92 bytes from 10.10.2.254: Destination Host Unreachable
> > > > 92 bytes from 10.10.2.254: Destination Host Unreachable
> > > > ^C--- 10.10.2.23 ping statistics ---
> > > >
> > > >
> > > >
> > > > Dchphost content
> > > >
> > > > # cat /etc/dhcphosts.txt
> > > > 02:00:72:36:00:01,10.10.2.83,tier12,infinite
> > > > 02:00:2f:16:00:0c,set:10_10_2_23,10.10.2.23,tier11,infinite
> > > > 02:00:78:24:00:01,10.10.1.51,tier11,infinite
> > > >
> > > >
> > > >
> > > > Dhcpopts content
> > > >
> > > >
> > > > # cat /etc/dhcpopts.txt
> > > > 10_10_2_23,3
> > > > 10_10_2_23,6
> > > > 10_10_2_23,15
> > > >
> > > >
> > > > I don't see any errors in /var/log/cloud.log on VR when secondary nic
> > is
> > > > added.
> > > > Is there any other I can troubleshoot this or look for logs?
> > > >
> > > > --
> > > > Thanks and regards
> > > > Rakesh
> > > >
> > >
> >
> >
> > --
> > Thanks and regards
> > Rakesh venkatesh
> >
>


-- 
Thanks and regards
Rakesh venkatesh


Re: Unable to ping secondary NIC ip of a VM in VPC

2020-08-10 Thread Rakesh Venkatesh
Hello Thomas


Thanks for the reply. It looks like the NIC is not configured properly in
VM. I dont see IP 10.10.2.23 being configured

# ip a
1: lo:  mtu 65536 qdisc noqueue state UNKNOWN group
default qlen 1
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
   valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
   valid_lft forever preferred_lft forever
2: eth0:  mtu 1500 qdisc pfifo_fast state
UP group default qlen 1000
link/ether 02:00:78:24:00:01 brd ff:ff:ff:ff:ff:ff
inet 10.10.1.51/24 brd 10.10.1.255 scope global eth0
   valid_lft forever preferred_lft forever
inet6 fe80::78ff:fe24:1/64 scope link
   valid_lft forever preferred_lft forever
3: eth1:  mtu 1500 qdisc noop state DOWN group default
qlen 1000
link/ether 02:00:1f:ed:00:0e brd ff:ff:ff:ff:ff:ff



I removed the secondary NIC, stopped the vm, added secondary NIC and
started the VM. The extra NIC is still not configured in VM.
Is this a bug?

On Mon, Aug 10, 2020 at 4:46 PM Thomas Joseph 
wrote:

> Hello Venkatesh,
>
> I'm guessing from the output provided that you have tried to ping the ip
> from the VR.
> Have you checked if the NIC is up on the VM?
>  you can run tcpdump on the related VM to check if the icmp packets are
> reaching the VM.
>
> Regards
> Thomas
>
> On Mon, 10 Aug 2020, 8:06 pm Rakesh Venkatesh, 
> wrote:
>
> > Hello,
> >
> > I have created a VPC with three tiers in it (tier1, tier2. tier3) using
> > cloudstack 4.14.0.0 version with Ubuntu16 as OS for mgt, agent
> > I created a VM in tier1 and I am able to ping its IP address
> >
> > # ping 10.10.1.51
> > PING 10.10.1.51 (10.10.1.51): 56 data bytes
> > 64 bytes from 10.10.1.51: icmp_seq=0 ttl=64 time=2.273 ms
> > 64 bytes from 10.10.1.51: icmp_seq=1 ttl=64 time=1.690 ms
> > ^C--- 10.10.1.51 ping statistics ---
> >
> >
> > After adding secondary NIC to vm from tier2 or tier3 then I am not able
> to
> > ping that IP address.
> >
> > # ping 10.10.2.23
> > PING 10.10.2.23 (10.10.2.23): 56 data bytes
> > 92 bytes from 10.10.2.254: Destination Host Unreachable
> > 92 bytes from 10.10.2.254: Destination Host Unreachable
> > 92 bytes from 10.10.2.254: Destination Host Unreachable
> > 92 bytes from 10.10.2.254: Destination Host Unreachable
> > ^C--- 10.10.2.23 ping statistics ---
> >
> >
> >
> > Dchphost content
> >
> > # cat /etc/dhcphosts.txt
> > 02:00:72:36:00:01,10.10.2.83,tier12,infinite
> > 02:00:2f:16:00:0c,set:10_10_2_23,10.10.2.23,tier11,infinite
> > 02:00:78:24:00:01,10.10.1.51,tier11,infinite
> >
> >
> >
> > Dhcpopts content
> >
> >
> > # cat /etc/dhcpopts.txt
> > 10_10_2_23,3
> > 10_10_2_23,6
> > 10_10_2_23,15
> >
> >
> > I don't see any errors in /var/log/cloud.log on VR when secondary nic is
> > added.
> > Is there any other I can troubleshoot this or look for logs?
> >
> > --
> > Thanks and regards
> > Rakesh
> >
>


-- 
Thanks and regards
Rakesh venkatesh


Unable to ping secondary NIC ip of a VM in VPC

2020-08-10 Thread Rakesh Venkatesh
Hello,

I have created a VPC with three tiers in it (tier1, tier2. tier3) using
cloudstack 4.14.0.0 version with Ubuntu16 as OS for mgt, agent
I created a VM in tier1 and I am able to ping its IP address

# ping 10.10.1.51
PING 10.10.1.51 (10.10.1.51): 56 data bytes
64 bytes from 10.10.1.51: icmp_seq=0 ttl=64 time=2.273 ms
64 bytes from 10.10.1.51: icmp_seq=1 ttl=64 time=1.690 ms
^C--- 10.10.1.51 ping statistics ---


After adding secondary NIC to vm from tier2 or tier3 then I am not able to
ping that IP address.

# ping 10.10.2.23
PING 10.10.2.23 (10.10.2.23): 56 data bytes
92 bytes from 10.10.2.254: Destination Host Unreachable
92 bytes from 10.10.2.254: Destination Host Unreachable
92 bytes from 10.10.2.254: Destination Host Unreachable
92 bytes from 10.10.2.254: Destination Host Unreachable
^C--- 10.10.2.23 ping statistics ---



Dchphost content

# cat /etc/dhcphosts.txt
02:00:72:36:00:01,10.10.2.83,tier12,infinite
02:00:2f:16:00:0c,set:10_10_2_23,10.10.2.23,tier11,infinite
02:00:78:24:00:01,10.10.1.51,tier11,infinite



Dhcpopts content


# cat /etc/dhcpopts.txt
10_10_2_23,3
10_10_2_23,6
10_10_2_23,15


I don't see any errors in /var/log/cloud.log on VR when secondary nic is
added.
Is there any other I can troubleshoot this or look for logs?

-- 
Thanks and regards
Rakesh


Re: Temporarily disabling API requests

2020-01-09 Thread Rakesh Venkatesh
Thanks Andrija. Yes that's the better way

On Thu, Jan 9, 2020 at 10:13 AM Andrija Panic 
wrote:

> You can try editing the User role (ACS 4.9+) and explicitly add the deny
> "ACL" for those API calls which you want to deny
> Later just remove the "ACL" rules.
>
> On Thu, 9 Jan 2020 at 09:59, Rakesh Venkatesh 
> wrote:
>
> > Hello folks
> >
> >
> > Is there a better way to disable all the VM related API's for a short
> > duration of time? Like I want to disable stop/start/migrate/destroy VM
> API.
> > I'm thinking of disabling the user so that the user cant send any API
> > requests. After some time I will enable the user back. Is there any other
> > better way to do it?
> >
> > --
> > Thanks and regards
> > Rakesh
> >
>
>
> --
>
> Andrija Panić
>


-- 
Thanks and regards
Rakesh venkatesh


Temporarily disabling API requests

2020-01-09 Thread Rakesh Venkatesh
Hello folks


Is there a better way to disable all the VM related API's for a short
duration of time? Like I want to disable stop/start/migrate/destroy VM API.
I'm thinking of disabling the user so that the user cant send any API
requests. After some time I will enable the user back. Is there any other
better way to do it?

-- 
Thanks and regards
Rakesh


Number of hosts in a cluster

2019-10-30 Thread Rakesh Venkatesh
Hello


In the documentation, I see that there is a configuration requirement of
not adding more than 16 hosts in a cluster: "For KVM, do not put more than
16 hosts in a cluster."

In our setup, we have more than 18 nodes in a cluster without any issues so
far. So I wanted to know what is the actual limit of hosts in a cluster. I
didn't see any such value in global setting as well. If there is no such
limitation then we need to update the documentation.

-- 
Thanks and regards
Rakesh venkatesh


Re: Virtual machines volume lock manager

2019-10-30 Thread Rakesh Venkatesh
Hi Andrija


Sorry for the late reply.

Im using 4.7 version of ACS. Qemu version 1:2.5+dfsg-5ubuntu10.40

Im not sure if ACS job failed or libvirt job as I didnt see into logs.
Yes the vm will be in paused state during migration but after the failed
migration, the same vm was in "running" state on two different hypervisors.
We wrote a script to find out how duplicated vm's are running and found out
that more than 5 vm's had this issue.


On Mon, Oct 28, 2019 at 2:42 PM Andrija Panic 
wrote:

> I've been running KVM public cloud up to recently and have never seen such
> behaviour.
>
> What versions (ACS, qemu, libvrit) are you running?
>
> How does the migration fail - ACS job - or libvirt job?
> destination VM is by default always in PAUSED state, until the migration is
> finished - only then the destination VM (on the new host) will get RUNNING,
> while previously pausing the original VM (on the old host).
>
> i,e.
> phase1  source vm RUNNING, destination vm PAUSED (RAM content being
> copied over... takes time...)
> phase2  source vm PAUSED, destination vm PAUSED (last bits of RAM
> content are migrated)
> phase3  source vm destroyed, destination VM RUNNING.
>
> Andrija
>
> On Mon, 28 Oct 2019 at 14:26, Rakesh Venkatesh 
> wrote:
>
> > Hello Users
> >
> >
> > Recently we have seen cases where when the Vm migration fails, cloudstack
> > ends up running two instances of the same VM on different hypervisors.
> The
> > state will be "running" and not any other transition state. This will of
> > course lead to corruption of disk. Does CloudStack has any option of
> volume
> > locking so that two instances of the same VM wont be running?
> > Anyone else has faced this issue and found some solution to fix it?
> >
> > We are thinking of using "virtlockd" of libvirt or implementing custom
> lock
> > mechanisms. There are some pros and cons of the both the solutions and i
> > want your feedback before proceeding further.
> >
> > --
> > Thanks and regards
> > Rakesh venkatesh
> >
>
>
> --
>
> Andrija Panić
>


-- 
Thanks and regards
Rakesh venkatesh


Virtual machines volume lock manager

2019-10-28 Thread Rakesh Venkatesh
Hello Users


Recently we have seen cases where when the Vm migration fails, cloudstack
ends up running two instances of the same VM on different hypervisors. The
state will be "running" and not any other transition state. This will of
course lead to corruption of disk. Does CloudStack has any option of volume
locking so that two instances of the same VM wont be running?
Anyone else has faced this issue and found some solution to fix it?

We are thinking of using "virtlockd" of libvirt or implementing custom lock
mechanisms. There are some pros and cons of the both the solutions and i
want your feedback before proceeding further.

-- 
Thanks and regards
Rakesh venkatesh


Re: Dedicated hosts for Domain/Account

2019-08-12 Thread Rakesh Venkatesh
Thanks for the quick reply.
I was browsing through the code and found the following


// check affinity group of type Explicit dedication exists. If No
put
// dedicated pod/cluster/host in avoid list
List vmGroupMappings =
_affinityGroupVMMapDao.findByVmIdType(vm.getId(), "ExplicitDedication");

if (vmGroupMappings != null && !vmGroupMappings.isEmpty()) {
isExplicit = true;
}


So this feature will work only if vm's are associated with affinity groups.
I created two vm's with same affinity group and after enabling the
maintenance mode they were migrated to the other dedicated machines.
So no need to create a github issue I guess.

On Mon, Aug 12, 2019 at 5:04 PM Andrija Panic 
wrote:

> Considering that manual VM LIVE migrations via CloudStack from
> non-dedicated to dedicated SHOULD/DOES work - then I would say this is an
> "unhandled" case, which indeed should be handled and live migration should
> happen instead of stopping the VMs.
>
> I assume someone else might jump in - but if not, please raise GitHub
> issues as a bug report.
>
>
> Thx
>
> On Mon, 12 Aug 2019 at 16:52, Rakesh Venkatesh 
> wrote:
>
> > Hello
> >
> > In my cloudstack setup, I have three KVM hypervisors out of which two
> > hypervisors are dedicated to Root/admin account and the third is not
> > dedicated. When I enable the maintenance mode on the dedicated
> hypervisor,
> > it will always migrate the vm's from dedicated to non dedicated
> hypervisor
> > but not to second dedicated hypervisor. I dont think this is the expected
> > behavior. Can any one please verify? The dedicated hypervisors will be
> > added to avoid set and the deployment planning manager skips these
> > hypervisors.
> >
> > If I dedicate the third hypervisor to different domain and enable the
> > maintenance mode on the first hypervisor then all the vm's will be
> stopped
> > instead of migrating to second dedicated hypervisor of the same
> > domain/account.
> >
> >
> > I have highlighted the necessary logs in red. You can see from the logs
> > that host with id 17 and 20 are dedicated but not 26. When maintenance
> mode
> > is enabled on host id 20, it skips 17 and 20 and migrates vm's to host id
> > 26
> >
> >
> >
> > 2019-08-12 14:35:23,754 DEBUG [c.c.d.DeploymentPlanningManagerImpl]
> > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c)
> > (logid:a16d7711) Deploy avoids pods: null, clusters: null, hosts: [20],
> > pools: null
> > 2019-08-12 14:35:23,757 DEBUG [c.c.d.DeploymentPlanningManagerImpl]
> > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c)
> > (logid:a16d7711) DeploymentPlanner allocation algorithm:
> > com.cloud.deploy.FirstFitPlanner@6fecace4
> > 2019-08-12 14:35:23,757 DEBUG [c.c.d.DeploymentPlanningManagerImpl]
> > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c)
> > (logid:a16d7711) Trying to allocate a host and storage pools from dc:8,
> > pod:8,cluster:null, requested cpu: 16000, requested ram: 8589934592
> > 2019-08-12 14:35:23,757 DEBUG [c.c.d.DeploymentPlanningManagerImpl]
> > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c)
> > (logid:a16d7711) Is ROOT volume READY (pool already allocated)?: Yes
> > 2019-08-12 14:35:23,757 DEBUG [c.c.d.DeploymentPlanningManagerImpl]
> > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c)
> > (logid:a16d7711) This VM has last host_id specified, trying to choose the
> > same host: 20
> > 2019-08-12 14:35:23,759 DEBUG [c.c.d.DeploymentPlanningManagerImpl]
> > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c)
> > (logid:a16d7711) The last host of this VM is in avoid set
> > 2019-08-12 14:35:23,759 DEBUG [c.c.d.DeploymentPlanningManagerImpl]
> > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c)
> > (logid:a16d7711) Cannot choose the last host to deploy this VM
> > 2019-08-12 14:35:23,759 DEBUG [c.c.d.FirstFitPlanner]
> > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c)
> > (logid:a16d7711) Searching resources only under specified Pod: 8
> > 2019-08-12 14:35:23,759 DEBUG [c.c.d.FirstFitPlanner]
> > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c)
> > (logid:a16d7711) Listing clusters in order of aggregate capacity, that
> have
> > (atleast one host with) enough CPU and RAM capacity under this Pod: 8
> > 2019-08-12 14:35:23,761 DEBUG [c.c.d.DeploymentPlanningManagerImpl]
> > (Work-Job-Executor-7:ctx-9f4363d1 job-473/job-246899 ctx-cef9b496)
> >

Dedicated hosts for Domain/Account

2019-08-12 Thread Rakesh Venkatesh
-246119/job-246902 ctx-9dbb7241)
(logid:b7e8e3a2) Is ROOT volume READY (pool already allocated)?: Yes
2019-08-12 14:35:23,766 DEBUG [c.c.d.DeploymentPlanningManagerImpl]
(Work-Job-Executor-8:ctx-1cc07ab1 job-246119/job-246902 ctx-9dbb7241)
(logid:b7e8e3a2) This VM has last host_id specified, trying to choose the
same host: 26
2019-08-12 14:35:23,780 DEBUG [c.c.d.DeploymentPlanningManagerImpl]
(Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c)
(logid:a16d7711) Checking resources in Cluster: 8 under Pod: 8
2019-08-12 14:35:23,782 DEBUG [c.c.c.CapacityManagerImpl]
(Work-Job-Executor-7:ctx-9f4363d1 job-473/job-246899 ctx-cef9b496)
(logid:bbb870bf) Host: 26 has cpu capability (cpu:48, speed:2900) to
support requested CPU: 1 and requested speed: 500
2019-08-12 14:35:23,782 DEBUG [c.c.c.CapacityManagerImpl]
(Work-Job-Executor-7:ctx-9f4363d1 job-473/job-246899 ctx-cef9b496)
(logid:bbb870bf) Checking if host: 26 has enough capacity for requested
CPU: 500 and requested RAM: 536870912 , cpuOverprovisioningFactor: 1.0
2019-08-12 14:35:23,782 DEBUG [c.c.a.m.a.i.FirstFitAllocator]
(Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c
FirstFitRoutingAllocator) (logid:a16d7711) Looking for hosts in dc: 8
 pod:8  cluster:8

-- 
Thanks and regards
Rakesh venkatesh


Querying async job result

2019-08-08 Thread Rakesh Venkatesh
Hello


I want to know what is the best way to query the async job result using
queryAsyncJobResult api. According to the documentation in
http://docs.cloudstack.apache.org/projects/archived-cloudstack-getting-started/en/latest/dev.html
,
the "jobstatus" of 1 means the command completed but im facing an issue
where even though the command is still running, the "jobstatus" is always 1.

Im running "prepareHostForMaintenance" command which returns the jobid.
When I run queryAsyncJobResult for this jobid, the jobstatus will always be
1 even though the hypervisor is still not in maintenance mode.

So can anyone tell me what is the best way to check if the hypervisor is in
maintenance mode or not? Im using 4.11 version


Below are the result which I get


"resourcestate": "PrepareForMaintenance",
"jobresultcode": 0,
  "jobresulttype": "object",
  "jobstatus": 1,

-- 
Thanks and regards
Rakesh venkatesh


Secondary Storage VM timeout issue every hour

2019-07-25 Thread Rakesh Venkatesh
 to Host 183 timed out after 3600
2019-07-25 04:01:22,785 DEBUG [c.c.a.m.AgentManagerImpl]
(AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to
listener: com.cloud.network.router.VpcVirtualNetworkApplianceManagerImpl
2019-07-25 04:01:22,785 DEBUG [c.c.a.m.AgentManagerImpl]
(AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to
listener: com.cloud.storage.download.DownloadListener




2019-07-25 04:01:22,785 ERROR [c.c.s.StatsCollector]
(StatsCollector-2:ctx-b55657a9) (logid:dafc4881) Error trying to retrieve
storage stats
com.cloud.utils.exception.CloudRuntimeException: Failed to send command,
due to Agent:183, com.cloud.exception.OperationTimedoutException: Commands
7541559051008607242 to Host 183 timed out after 3600
at
org.apache.cloudstack.storage.RemoteHostEndPoint.sendMessage(RemoteHostEndPoint.java:133)
at
com.cloud.server.StatsCollector$StorageCollector.runInContext(StatsCollector.java:1139)
at
org.apache.cloudstack.managed.context.ManagedContextRunnable$1.run(ManagedContextRunnable.java:49)
at
org.apache.cloudstack.managed.context.impl.DefaultManagedContext$1.call(DefaultManagedContext.java:56)
at
org.apache.cloudstack.managed.context.impl.DefaultManagedContext.callWithContext(DefaultManagedContext.java:103)
at
org.apache.cloudstack.managed.context.impl.DefaultManagedContext.runWithContext(DefaultManagedContext.java:53)
at
org.apache.cloudstack.managed.context.ManagedContextRunnable.run(ManagedContextRunnable.java:46)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
2019-07-25 04:01:22,786 DEBUG [c.c.a.m.AgentManagerImpl]
(AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to
listener: com.cloud.consoleproxy.ConsoleProxyListener
2019-07-25 04:01:22,789 DEBUG [c.c.a.m.AgentManagerImpl]
(AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to
listener: com.cloud.storage.LocalStoragePoolListener
2019-07-25 04:01:22,789 DEBUG [c.c.a.m.AgentManagerImpl]
(AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to
listener: com.cloud.storage.upload.UploadListener
2019-07-25 04:01:22,790 DEBUG [c.c.a.m.AgentManagerImpl]
(AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to
listener: com.cloud.capacity.StorageCapacityListener
2019-07-25 04:01:22,790 DEBUG [c.c.a.m.AgentManagerImpl]
(AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to
listener: com.cloud.capacity.ComputeCapacityListener
2019-07-25 04:01:22,790 DEBUG [c.c.a.m.AgentManagerImpl]
(AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to
listener: com.cloud.network.SshKeysDistriMonitor
2019-07-25 04:01:22,791 DEBUG [c.c.a.m.AgentManagerImpl]
(AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to
listener: com.cloud.network.router.VirtualNetworkApplianceManagerImpl
2019-07-25 04:01:22,791 DEBUG [c.c.a.m.AgentManagerImpl]
(AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to
listener:
com.cloud.network.NetworkUsageManagerImpl$DirectNetworkStatsListener
2019-07-25 04:01:22,791 DEBUG [c.c.n.NetworkUsageManagerImpl]
(AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Disconnected called on 183
with status Disconnected




2019-07-25 04:01:22,791 DEBUG [c.c.a.m.AgentManagerImpl]
(AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to
listener: com.cloud.agent.manager.AgentManagerImpl$BehindOnPingListener
2019-07-25 04:01:22,791 DEBUG [c.c.a.m.AgentManagerImpl]
(AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to
listener: com.cloud.agent.manager.AgentManagerImpl$SetHostParamsListener
2019-07-25 04:01:22,791 DEBUG [c.c.h.Status] (AgentTaskPool-1:ctx-66de2057)
(logid:841d2a63) Transition:[Resource state = Enabled, Agent event =
ShutdownRequested, Host id = 183, name = s-2775-VM]



-- 
Thanks and regards
Rakesh venkatesh


Get VM OS type

2019-06-17 Thread Rakesh Venkatesh
Hello Folks

Is there a way to know whether the VM is running on Windows or Linux OS? I
can't reply on OS type because we can use Ubuntu as OS type for Windows VM.
Even though the os type is Linux/Ubuntu, the VM is running on Windows
version. Another way is to open the console and see the login screen. This
will get the actual data but I want to do automation to see for all VM's
and opening the console is not feasible to automate. Is there any other way
to get it?

-- 
Thanks and regards
Rakesh Venkatesh


Switching backup VR to master manually

2019-05-20 Thread Rakesh Venkatesh
Hello

Is there a way to switch the backup virtual router to master state manually
by triggering something or running some scripts?

Currently when I restart network with cleanup option, first the backup will
be upgraded to newer cloudstack version and when the master is destroyed
then only backup becomes the master and this process might take atleast
5-10 seconds. In the meantime there will be dowmtime. Is there a way to
quickly change the state?

Im thinking on stopping the keeaplived on master so that backup will become
master quickly. Is there a better way to do it?

-- 
Thanks and regards
Rakesh venkatesh


Re: Troubleshooting the Exception "Unable to start VM instance"

2019-04-29 Thread Rakesh Venkatesh
Hello Peter

If you dig into logs a little bit more, you can see the actual reason for
the error. The usual errors are insufficient server capacity exception like
not enough cpu, memory, ip address or mismatch between template tag and
host tag or some other mismatch in the tags. These are the common errors I
have seen.
Search for "startVirtualMachine" api and trace its jobid which is usually
like "context-" and look at the logs

On Mon, Apr 29, 2019 at 12:58 PM  wrote:

> Hi, all,
>
> sometimes the system can't start a VM instance (a 530 error).
> In the logs there is the following trace - how to troubleshoot it and
> understand which circumstances lead to it?
>
>
> com.cloud.utils.exception.CloudRuntimeException: Unable to start VM
> instance
> at
> com.cloud.vm.VirtualMachineManagerImpl.orchestrateStart(VirtualMachineManagerImpl.java:4889)
> at sun.reflect.GeneratedMethodAccessor433.invoke(Unknown Source)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> com.cloud.vm.VmWorkJobHandlerProxy.handleVmWorkJob(VmWorkJobHandlerProxy.java:107)
> at
> com.cloud.vm.VirtualMachineManagerImpl.handleVmWorkJob(VirtualMachineManagerImpl.java:5047)
> at
> com.cloud.vm.VmWorkJobDispatcher.runJob(VmWorkJobDispatcher.java:102)
> at
> org.apache.cloudstack.framework.jobs.impl.AsyncJobManagerImpl$5.runInContext(AsyncJobManagerImpl.java:615)
> at
> org.apache.cloudstack.managed.context.ManagedContextRunnable$1.run(ManagedContextRunnable.java:49)
> at
> org.apache.cloudstack.managed.context.impl.DefaultManagedContext$1.call(DefaultManagedContext.java:56)
> at
> org.apache.cloudstack.managed.context.impl.DefaultManagedContext.callWithContext(DefaultManagedContext.java:103)
> at
> org.apache.cloudstack.managed.context.impl.DefaultManagedContext.runWithContext(DefaultManagedContext.java:53)
> at
> org.apache.cloudstack.managed.context.ManagedContextRunnable.run(ManagedContextRunnable.java:46)
> at
> org.apache.cloudstack.framework.jobs.impl.AsyncJobManagerImpl$5.run(AsyncJobManagerImpl.java:563)
> at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748)
>
> kind regards
> Peter
>
>

-- 
Thanks and regards
Rakesh venkatesh


Re: Cloudstack database timeout value

2019-04-11 Thread Rakesh Venkatesh
Thanks Suresh. I will try that out

On Wed, Apr 10, 2019 at 6:32 PM Suresh Kumar Anaparti <
sureshkumar.anapa...@gmail.com> wrote:

> Hi Rakesh,
>
> Locate "db.properties" file (may be at /etc/cloudstack/management/) in the
> management server and check for the parameters db.cloud.maxWait (default:
> 10 secs), db.cloud.autoReconnect (default: true) and other database
> settings. Restart the management server post parameters update.
>
> - Suresh
>
> On Wed, Apr 10, 2019 at 7:41 PM Rakesh Venkatesh <
> www.rakeshv@gmail.com>
> wrote:
>
> > Hello Folks
> >
> > Is there a way to increase the database connectivity timeout value in
> case
> > cloudstack loses connection to the database? Also is it possible to
> > configure the value? Can you point me to the right place? I also want to
> > increase the retries value so that it will retry to connect to DB
> > configured value of times and then really shuts down.
> >
> > We have a redundant setup with three database hosts and switching from
> one
> > db host to another db host took more than 10 seconds and mgt server got
> > killed since it cant connect to db anymore. i have to restart the mgt
> > server to bring it back up. To avoid this I want to configure the timeout
> > value as well as the number of retries to get a connection
> >
> > --
> > Thanks and regards
> > Rakesh venkatesh
> >
>


-- 
Thanks and regards
Rakesh venkatesh


Cloudstack database timeout value

2019-04-10 Thread Rakesh Venkatesh
Hello Folks

Is there a way to increase the database connectivity timeout value in case
cloudstack loses connection to the database? Also is it possible to
configure the value? Can you point me to the right place? I also want to
increase the retries value so that it will retry to connect to DB
configured value of times and then really shuts down.

We have a redundant setup with three database hosts and switching from one
db host to another db host took more than 10 seconds and mgt server got
killed since it cant connect to db anymore. i have to restart the mgt
server to bring it back up. To avoid this I want to configure the timeout
value as well as the number of retries to get a connection

-- 
Thanks and regards
Rakesh venkatesh


Re: Shutdown VM

2019-04-05 Thread Rakesh Venkatesh
Disable the HA for the vm by changing filed in vm_instance table and
shutdown the vm. The cloudstack wont restart it. That worked for me

On Fri, Apr 5, 2019 at 1:25 PM Alex Adati  wrote:

> Hello,
>
> I'm using a computational offering with HA enabled, and would like to know
> if there is any way to shut down the VM by the operating system.
>
> Thank's.
>
> Em sex, 5 de abr de 2019 às 04:54, Riepl, Gregor (SWISS TXT) <
> gregor.ri...@swisstxt.ch> escreveu:
>
> > > You should be using a computer offering with HA enabled.
> > > When you’re using this option, cloudstack will start vm until you
> > shutdown it by cloudstack interface (UI or api).
> >
> > I think you meant: You should NOT be using a compute offering with HA
> > enabled, right?
> >
>
>
> --
>


-- 
Thanks and regards
Rakesh venkatesh


Re: Network metrics

2019-04-05 Thread Rakesh Venkatesh
Hello Gregor


You can use "libvirt exporter" to export all stats like cpu utilization,
memory consumption, disk read/write, network read/write and export it to
prometheus. From prometheus you can use grafana or other monitoring tools
to see all stats. We use exporter merger third party tool to merge all the
data from all the hypervisors and show it on dashboard. This way if any of
your vm is having issues, you can quickly look at graphs and troubleshoot
it.

On Fri, Apr 5, 2019 at 10:01 AM Riepl, Gregor (SWISS TXT) <
gregor.ri...@swisstxt.ch> wrote:

> Hi,
>
> I recently discovered that CloudStack exposes certain VM metrics like disk
> I/O, network bandwidth, CPU load and (if the HV agent is installed) memory
> usage.
> Presumably, these are fetched directly from the hypervisor.
>
> However: I found no way to get the same statistics for system VMs, in
> particular virtual routers.
> Does the CloudStack API offer a way to obtain per-network metrics?
>
> Regards,
> Gregor



-- 
Thanks and regards
Rakesh venkatesh


Calling cloudstack container service API's from Cloudstack

2019-01-21 Thread Rakesh Venkatesh
Hello Folks


I am wondering if there is a way to call cloudstack container service API
from cloudstack? I want to delete all the containers created by a domain
when the domain is deleted. I am using cloudstack version 4.7 with KVM and
when I delete a domain, the containers created by domain are still active.
So I want to call deleteContainerCluster API from cloudstack deleteDomain
API.

What modifications should I make
in server/src/com/cloud/user/DomainManagerImpl.java of Cloudstack or
DeleteContainerClusterCmd.java of CCS? Any examples of this usage are
helpful.

-- 
Thanks and regards
Rakesh venkatesh


Re: Both routers are in Master state in VPC

2018-12-06 Thread Rakesh Venkatesh
Hi Rohit


Thanks for the reply. We saw this issue in our customer setup. we are using
kvm hypervisor and cloudstack 4.7.1
I will have a look at the issue you provided.

On Thu, Dec 6, 2018 at 3:28 PM Rohit Yadav 
wrote:

> Hi Rakesh,
>
>
> Can you share your configuration/setup of VPC, hypervisor, are you using
> private gateways and/or VPN with your VPCs and the version of CloudStack?
>
>
> There are few known cases where both VPC redundant VRs become master, see:
> https://github.com/apache/cloudstack/pull/2128 and CLOUDSTACK-9885 issue.
>
>
> - Rohit
>
> <https://cloudstack.apache.org>
>
>
>
> 
> From: Rakesh Venkatesh 
> Sent: Thursday, December 6, 2018 4:53:31 PM
> To: users@cloudstack.apache.org
> Subject: Both routers are in Master state in VPC
>
> Hello Folks
>
> Several times we have encountered that both the master and the backup
> router of VPC network will be in MASTER state and this causes several
> issues to network connectivity. We dont the trigger for it but can be
> solved by restarting the network with cleanup option. Anyone else has
> encountered this issue before and know how to fix it?
>
> rohit.ya...@shapeblue.com
> www.shapeblue.com
> Amadeus House, Floral Street, London  WC2E 9DPUK
> @shapeblue
>
>
>
>

-- 
Thanks and regards
Rakesh venkatesh


Both routers are in Master state in VPC

2018-12-06 Thread Rakesh Venkatesh
Hello Folks

Several times we have encountered that both the master and the backup
router of VPC network will be in MASTER state and this causes several
issues to network connectivity. We dont the trigger for it but can be
solved by restarting the network with cleanup option. Anyone else has
encountered this issue before and know how to fix it?


Re: VRs swapping with 256 MB RAM

2018-09-20 Thread Rakesh Venkatesh
Hello Rene

Even for VR's running on KVM, 256MB is really less. Thats why we offer
extra VR with 2 cores and 1GB RAM as another option so that customers can
use it instead of the default 256MB and 1 core.

On Tue, Sep 18, 2018 at 5:56 PM Rene Moser  wrote:

> Hi
>
> While running test for a 4.11.1 (VMware) upgrade in our lab, we run into
> low memory / swapping of VRs having 256 MB RAM. After 2-3 days it became
> critical because the management server connections to VRs took very
> long, minutes, this resulted in many more problems all over.
>
> Make sure your VRs have enough RAM.
>
> Regards
> René
>


-- 
Thanks and regards
Rakesh venkatesh


Re: Rename an instance

2018-09-20 Thread Rakesh Venkatesh
Hello Ugo

You can change the display name of a vm through database or cloudmonkey
without stopping the vm. You need to change the name in two tables if you
are doing through database

Change display_name in "vm_instance" and "user_vm" tables. We did it for
our customers and didnt face any issue

On Wed, Sep 19, 2018 at 12:28 PM Ugo Vasi 
wrote:

> Hi all,
> from the cloudstack web interface (version 4.11.1 in my case) I can only
> change the "Display name" by stopping the instance and clicking the icon.
>
> Is it possible to change the instance name?
>
> Is it dangerous to change the field in the database?
>
> Thanks in advance
> --
>
> *Ugo Vasi* / System Administrator
> ugo.v...@procne.it <mailto:ugo.v...@procne.it>
>
>
>
>
> *Procne S.r.l.*
> +39 0432 486 523
> via Cotonificio, 45
> 33010 Tavagnacco (UD)
> www.procne.it <http://www.procne.it/>
>
>
> Le informazioni contenute nella presente comunicazione ed i relativi
> allegati possono essere riservate e sono, comunque, destinate
> esclusivamente alle persone od alla Società sopraindicati. La
> diffusione, distribuzione e/o copiatura del documento trasmesso da parte
> di qualsiasi soggetto diverso dal destinatario è proibita sia ai sensi
> dell'art. 616 c.p., che ai sensi del Decreto Legislativo n. 196/2003
> "Codice in materia di protezione dei dati personali". Se avete ricevuto
> questo messaggio per errore, vi preghiamo di distruggerlo e di informare
> immediatamente Procne S.r.l. scrivendo all' indirizzo e-mail
> i...@procne.it <mailto:i...@procne.it>.
>
>

-- 
Thanks and regards
Rakesh venkatesh


Re: Disk controller type

2018-09-20 Thread Rakesh Venkatesh
I think you can change the controller type in XML file of the VM but not
sure if that works or not.

virsh dumpxml 
virsh edit 
change the controller type, save and start the vm again

On Thu, Sep 20, 2018 at 9:59 AM Adam Witwicki 
wrote:

> Hello,
>
> I have a number of of instances using the wrong controller type (ATA) does
> anyone know where this is set in the database so I can start them with the
> VirtIO scsi controller?
>
> Thanks
>
> Adam
>
>
>
>
> Disclaimer Notice:
> This email has been sent by Oakford Technology Limited, while we have
> checked this e-mail and any attachments for viruses, we can not guarantee
> that they are virus-free. You must therefore take full responsibility for
> virus checking.
> This message and any attachments are confidential and should only be read
> by those to whom they are addressed. If you are not the intended recipient,
> please contact us, delete the message from your computer and destroy any
> copies. Any distribution or copying without our prior permission is
> prohibited.
> Internet communications are not always secure and therefore Oakford
> Technology Limited does not accept legal responsibility for this message.
> The recipient is responsible for verifying its authenticity before acting
> on the contents. Any views or opinions presented are solely those of the
> author and do not necessarily represent those of Oakford Technology Limited.
> Registered address: Oakford Technology Limited, 10 Prince Maurice Court,
> Devizes, Wiltshire. SN10 2RT.
> Registered in England and Wales No. 5971519
>
>

-- 
Thanks and regards
Rakesh venkatesh


Re: Error while migrating vm in ACS 4.7.1

2018-08-10 Thread Rakesh Venkatesh
Hello Dag

I have seen this issue in two scenarios


1. While migrating multiple vm's
2. When a host is put into maintenance mode


So along with this vm there are other vm's that are being migrated as well.

I also see this in the log which is confusing me

|2018-08-10 12:17:26,030 INFO
[resource.wrapper.LibvirtMigrateCommandWrapper]
(agentRequest-Handler-4:null) Migration thread for i-884-14365-VM is done


Does that answer your question?

On Fri, Aug 10, 2018 at 4:21 PM Dag Sonstebo 
wrote:

> Hi Rakesh,
>
> “There's a migration process in progress” > have you done some digging to
> find out what is actually running?
>
> Regards,
> Dag Sonstebo
> Cloud Architect
> ShapeBlue
>
>
> dag.sonst...@shapeblue.com
> www.shapeblue.com
> 53 Chandos Place, Covent Garden, London  WC2N 4HSUK
> @shapeblue
>
>
>
> On 10/08/2018, 15:16, "Rakesh Venkatesh" 
> wrote:
>
> Hello Folks
>
> We are seeing some errors consistently while migrating vm's from one
> node
> another node but still not able to find out the root cause of it. Can
> anyone please let me know what might be causing the issue? The vm's are
> migrated successfully though
>
> This issue happens when I try to migrate the vm either from cloudmonkey
> command or through UI.
>
> Management server:
>
> $ lsb_release -a
> No LSB modules are available.
> Distributor ID: Ubuntu
> Description: Ubuntu 12.04.3 LTS
> Release: 12.04
> Codename: precise
>
>
> Error logs from management server:
>
> Aug 10 12:35:48 node13 sh[50956]: 2018-08-10 12:35:48,243{GMT} INFO [
> resource.wrapper.LibvirtMigrateCommandWrapper]
> (agentRequest-Handler-5:)
> Waiting for migration of i-884-15037-VM to complete, waited 3000ms
> Aug 10 12:35:48 node13 sh[50956]: 2018-08-10 12:35:48,486{GMT} INFO [
> resource.wrapper.LibvirtMigrateCommandWrapper]
> (agentRequest-Handler-1:)
> Waiting for migration of i-884-14365-VM to complete, waited 1ms
> Aug 10 12:35:49 node13 sh[50956]: libvirt: QEMU Driver error : internal
> error: unable to execute QEMU command 'migrate-set-capabilities':
> There's a
> migration process in progressAug 10 12:35:49 node13 sh[50956]:
> 2018-08-10
> 12:35:49,087{GMT} INFO [resource.wrapper.LibvirtMigrateCommandWrapper]
> (agentRequest-Handler-1:) Migration thread for i-884-14365-VM is done
> Aug 10 12:35:49 node13 sh[50956]: 2018-08-10 12:35:49,087{GMT} DEBUG
> [resource.wrapper.LibvirtMigrateCommandWrapper]
> (agentRequest-Handler-1:)
> Failed to execute while migrating domain: org.libvirt.LibvirtException:
> internal error: unable to execute QEMU command
> 'migrate-set-capabilities':
> There's a migration process in progress
> Aug 10 12:35:49 node13 sh[50956]: 2018-08-10 12:35:49,087{GMT} DEBUG
> [kvm.
> resource.LibvirtComputingResource] (agentRequest-Handler-1:) Restoring
> the
> metadata of vm snapshots of vm i-884-14365-VM
> Aug 10 12:35:49 node13 sh[50956]: 2018-08-10 12:35:49,088{GMT} DEBUG
> [cloud.
> agent.Agent] (agentRequest-Handler-1:) Seq 367-5848768541070742315: {
> Ans: ,
> MgmtId: 48674822042491, via: 367, Ver: v1, Flags: 10, [{
>
> "com.cloud.agent.api.MigrateAnswer":{"result":false,"details":"org.libvirt.LibvirtException:
> internal error: unable to execute QEMU command
> 'migrate-set-capabilities':
> There's a migration process in progress","wait":0}}] }
>
>
> Agent:
>
> # lsb_release -a No LSB modules are available. Distributor ID: Ubuntu
> Description: Ubuntu 16.04.3 LTS Release: 16.04 Codename: xenial
>
> Error logs from agent:
>
>
> 2018-08-10 12:17:26,030 DEBUG
> [resource.wrapper.LibvirtMigrateCommandWrapper]
> (agentRequest-Handler-4:null) Failed to execute while migrating domain:
> org.libvirt.LibvirtException: internal error: unable to execute QEMU
> command 'migrate-set-capabilities': There's a migration process in
> progress
> 2018-08-10 12:17:26,030 DEBUG [kvm.resource.LibvirtComputingResource]
> (agentRequest-Handler-4:null) Restoring the metadata of vm snapshots
> of vm
> i-884-14365-VM 2018-08-10 12:17:26,031 DEBUG [cloud.agent.Agent]
> (agentRequest-Handler-4:null) Seq 367-5848768541070742169: { Ans: ,
> MgmtId:
> 48674822042491, via: 367, Ver: v1, Flags: 10,
>
> [{"com.cloud.agent.api.MigrateAnswer":{"result":false,"details":"org.libvirt.LibvirtException:
> internal error: unable to execute QEMU command
> 'migrate-set-capabilities':
> There's a migration process in progress","wait":0}}] }
>
> 2018-08-10 12:17:26,564 INFO
> [resource.wrapper.LibvirtMigrateCommandWrapper]
> (agentRequest-Handler-4:null) Live migration of instance i-884-14365-VM
> initiated
>
>
> --
> Thanks and regards
> Rakesh venkatesh
>
>
>

-- 
Thanks and regards
Rakesh venkatesh


Error while migrating vm in ACS 4.7.1

2018-08-10 Thread Rakesh Venkatesh
Hello Folks

We are seeing some errors consistently while migrating vm's from one node
another node but still not able to find out the root cause of it. Can
anyone please let me know what might be causing the issue? The vm's are
migrated successfully though

This issue happens when I try to migrate the vm either from cloudmonkey
command or through UI.

Management server:

$ lsb_release -a
No LSB modules are available.
Distributor ID: Ubuntu
Description: Ubuntu 12.04.3 LTS
Release: 12.04
Codename: precise


Error logs from management server:

Aug 10 12:35:48 node13 sh[50956]: 2018-08-10 12:35:48,243{GMT} INFO [
resource.wrapper.LibvirtMigrateCommandWrapper] (agentRequest-Handler-5:)
Waiting for migration of i-884-15037-VM to complete, waited 3000ms
Aug 10 12:35:48 node13 sh[50956]: 2018-08-10 12:35:48,486{GMT} INFO [
resource.wrapper.LibvirtMigrateCommandWrapper] (agentRequest-Handler-1:)
Waiting for migration of i-884-14365-VM to complete, waited 1ms
Aug 10 12:35:49 node13 sh[50956]: libvirt: QEMU Driver error : internal
error: unable to execute QEMU command 'migrate-set-capabilities': There's a
migration process in progressAug 10 12:35:49 node13 sh[50956]: 2018-08-10
12:35:49,087{GMT} INFO [resource.wrapper.LibvirtMigrateCommandWrapper]
(agentRequest-Handler-1:) Migration thread for i-884-14365-VM is done
Aug 10 12:35:49 node13 sh[50956]: 2018-08-10 12:35:49,087{GMT} DEBUG
[resource.wrapper.LibvirtMigrateCommandWrapper] (agentRequest-Handler-1:)
Failed to execute while migrating domain: org.libvirt.LibvirtException:
internal error: unable to execute QEMU command 'migrate-set-capabilities':
There's a migration process in progress
Aug 10 12:35:49 node13 sh[50956]: 2018-08-10 12:35:49,087{GMT} DEBUG [kvm.
resource.LibvirtComputingResource] (agentRequest-Handler-1:) Restoring the
metadata of vm snapshots of vm i-884-14365-VM
Aug 10 12:35:49 node13 sh[50956]: 2018-08-10 12:35:49,088{GMT} DEBUG [cloud.
agent.Agent] (agentRequest-Handler-1:) Seq 367-5848768541070742315: { Ans: ,
MgmtId: 48674822042491, via: 367, Ver: v1, Flags: 10, [{
"com.cloud.agent.api.MigrateAnswer":{"result":false,"details":"org.libvirt.LibvirtException:
internal error: unable to execute QEMU command 'migrate-set-capabilities':
There's a migration process in progress","wait":0}}] }


Agent:

# lsb_release -a No LSB modules are available. Distributor ID: Ubuntu
Description: Ubuntu 16.04.3 LTS Release: 16.04 Codename: xenial

Error logs from agent:


2018-08-10 12:17:26,030 DEBUG
[resource.wrapper.LibvirtMigrateCommandWrapper]
(agentRequest-Handler-4:null) Failed to execute while migrating domain:
org.libvirt.LibvirtException: internal error: unable to execute QEMU
command 'migrate-set-capabilities': There's a migration process in progress
2018-08-10 12:17:26,030 DEBUG [kvm.resource.LibvirtComputingResource]
(agentRequest-Handler-4:null) Restoring the metadata of vm snapshots of vm
i-884-14365-VM 2018-08-10 12:17:26,031 DEBUG [cloud.agent.Agent]
(agentRequest-Handler-4:null) Seq 367-5848768541070742169: { Ans: , MgmtId:
48674822042491, via: 367, Ver: v1, Flags: 10,
[{"com.cloud.agent.api.MigrateAnswer":{"result":false,"details":"org.libvirt.LibvirtException:
internal error: unable to execute QEMU command 'migrate-set-capabilities':
There's a migration process in progress","wait":0}}] }

2018-08-10 12:17:26,564 INFO
[resource.wrapper.LibvirtMigrateCommandWrapper]
(agentRequest-Handler-4:null) Live migration of instance i-884-14365-VM
initiated


-- 
Thanks and regards
Rakesh venkatesh