RCE found in log4j library
Hello users/devs Recently there was an RCE found in log4j library. Fortunately we are running on older version and so we are good I guess. But if we plan to upgrade in future, we need to keep the library version in mind https://www.lunasec.io/docs/blog/log4j-zero-day/ -- Thanks and regards Rakesh
How to use @Configuration annotation
Hello devs I am trying to integrate third party library into cloudstack but the library supports @Configuration and @Enable* annotations to use it and not with xml files. I tried adding that annotation on a config file, built a new package but the cloudstack wont boot. It is stuck at below line 2021-11-24 11:40:50,554 INFO [o.a.c.s.m.m.i.DefaultModuleDefinitionSet] (main:null) (logid:) Loading module context [core] from URL [jar:file:/usr/share/cloudstack-management/lib/cloudstack-4.16.1.0-SNAPSHOT.jar!/META-INF/cloudstack/system/spring-contrail-system-context-inheritable.xml] 2021-11-24 11:40:50,554 INFO [o.a.c.s.m.m.i.DefaultModuleDefinitionSet] (main:null) (logid:) Loading module context [core] from URL [jar:file:/usr/share/cloudstack-management/lib/cloudstack-4.16.1.0-SNAPSHOT.jar!/META-INF/cloudstack/bootstrap/spring-bootstrap-context-inheritable.xml] Is there anything else I need to do to enable this feature? I also added bean id definition in one of the xml files under server/ but that also didnt help. -- Thanks and regards Rakesh
api to import network acl rules
Hello I have a network acl which has around 800 rules. i want to create another network acl with these rules along with some extra rules. Is there a way to just copy all these 800 rules from one network acl to another one? I see the export acl's button but there is no option to import it. Is there any other simple way? -- Thanks and regards Rakesh
2FA
Hello Has anyone thought about 2FA or about how to implement it in cloudstack? Looks like this will be good addition to enhance the security. I have some idea about implementing in the backend but dont have much idea on how to display the QR code in ui or other functionalities which is needed for frontend part. -- Thanks and regards Rakesh
Re: Limiting the conntrackd logs in virtual router
Hello Andrija Sorry for replying to your email this late. Just saw it. Basically there was an issue with logrotate service as it was running just once a day instead of hourly. I have created a pr for that in 4.15 On Tue, Jun 15, 2021 at 12:41 AM Andrija Panic wrote: > Rakesh, > > can you confirm if you are seeing this for ALL VPCs with S-2-S enabled > can you describe/quantify the MB/s log increase over time - does it have > anything to do with the traffic throughput through the tunnel, or is it > just happening even with no traffic (i.e. you can setup S-2-S tunnel > between 2 VPCs (one of them marked as "passive") - I would like to see/test > this in 4.15.1 (if you confirm that the amount of traffic does NOT > influence the size of the logs) > > Please respond ASAP - as we need to craft 4.15.1 RC2 - if there is a bug, > we need to fix it (and I will help testing). > > Best, > > > On Wed, 9 Jun 2021 at 13:31, Rakesh Venkatesh > wrote: > > > If I restart with cleanup then the modified config will be deleted and > the > > log starts populating very fast again. > > > > On Wed, Jun 9, 2021 at 1:19 PM Andrija Panic > > wrote: > > > > > Did you try restarting that VPC with cleanup - so brand new VR is > > created - > > > to see if the issue is still observed. > > > > > > Can you test the same now with 4.15.1 RC1 (voting is happening, please > > feel > > > free to test and participate!) > > > > > > Best, > > > > > > On Wed, 9 Jun 2021 at 10:57, Rakesh Venkatesh < > www.rakeshv@gmail.com > > > > > > wrote: > > > > > > > I have 4.14 version. The issue exists with only 1 customer. > > > > The VR's belonging to other customers are fine but for one customer > who > > > has > > > > enabled site 2 site vpn, its creating lot of logs. > > > > The config I mentioned above is a temporary fix as it will be deleted > > > once > > > > the network is restarted with cleanup. > > > > > > > > On Wed, Jun 9, 2021 at 10:52 AM Andrija Panic < > andrija.pa...@gmail.com > > > > > > > wrote: > > > > > > > > > Can you advise which version and have you tried isolating the > > problem - > > > > > this is first time I hear for such an issue (and I have customers > > with > > > 3y > > > > > VR uptime with no issues) > > > > > > > > > > Perhaps something is broken inside that VR, thus the excessive > > logging, > > > > > etc? > > > > > > > > > > Best, > > > > > > > > > > On Wed, 9 Jun 2021 at 10:07, Rakesh Venkatesh < > > > www.rakeshv@gmail.com > > > > > > > > > > wrote: > > > > > > > > > > > Hello > > > > > > > > > > > > I have an issue in the virtual router where > > > conntrackd(conntrack-tools) > > > > > is > > > > > > populating /var/log/daemon.log a lot and because of that VR is > > > running > > > > > out > > > > > > of disk space. Do you guys know how to prevent conntrackd from > > > logging > > > > > too > > > > > > much? I manually changed the logrotate.conf to rotate every hour > > and > > > > > retain > > > > > > only few copies but thats a quick fix. Is there any better fix to > > > avoid > > > > > > this issue? In our setup the VR has 2GB disk and even that got > > filled > > > > up > > > > > > pretty quickly. > > > > > > > > > > > > The logrotate config for daemon.log is present in > > > > > /etc/logrotate.d/rsyslog > > > > > > and I added > > > > > > > > > > > > /var/log/daemon.log > > > > > > { > > > > > > hourly > > > > > > rotate 1 > > > > > > missingok > > > > > > notifempty > > > > > > compress > > > > > > copytruncate > > > > > > } > > > > > > > > > > > > > > > > > > The config present in /etc/logrotate.d/conntrackd is pointing > > > > > > to /var/log/conntrackd-stats.log and so changing that config wont > > > help. > > > > > > > > > > > > -- > > > > > > Thanks and regards > > > > > > Rakesh > > > > > > > > > > > > > > > > > > > > > -- > > > > > > > > > > Andrija Panić > > > > > > > > > > > > > > > > > -- > > > > Thanks and regards > > > > Rakesh venkatesh > > > > > > > > > > > > > -- > > > > > > Andrija Panić > > > > > > > > > -- > > Thanks and regards > > Rakesh venkatesh > > > > > -- > > Andrija Panić > -- Thanks and regards Rakesh venkatesh
Code executed twice while deploying VM
Hello Users/Devs Today I observed that while deploying a VM, the same code is executed twice: the first time, while trying to find a suitable deployment destination and the second time, while trying to start a VM. I think this is a redundant process and also time-consuming since the same calculations are done twice. These are the one of the duplicate logs Asyn job is created 2021-07-28 09:01:27,964 (logid:10ac9f28) ZoneWideStoragePoolAllocator to find storage pool Now it finds a suitable destination Returning Deployment Destination: Dest[Zone(Id)-Pod(Id)-Cluster(Id)-Host(Id)-Storage(Volume(Id|Type-->Pool(Id))] : Dest[Zone(1)-Pod(1)-Cluster(1)-Host(1)-Storage(Volume(23|ROOT-->Pool(3))] VM start attempt #1 2021-07-28 09:01:36,039 job-174/job-175 ctx-b815930d) (logid:10ac9f28) ZoneWideStoragePoolAllocator to find storage pool >From the above two logs, I can see that starting at 9:01:27 and 09:01:36, its doing the same calculations. if I see the logs before and after the above-mentioned logs, they are all same. They try to find a suitable host, a suitable storage pool, and so on. In the production platform, the calculation is done twice and this takes a lot of time to deploy a VM. Do you guys also think this is an issue or redundant process and needs to be improved? Or any other suggestions to avoid double calculation? -- Thanks and regards Rakesh
Disabling a storage pool
Hello folks Is there a way to disable a particular storage pool so that it won't be used for further volume allocation? I don't want to enable the maintenance mode as that will turn off the VM's whose volumes running on that pool. I don't want to use a global setting also since this will come into effect after the threshold value is reached. In some cases even if the pool is just 10% allocated, I still want to disable it so that the current volumes will keep existing on the same pool and at the same time further deployment of volumes on this pool is disabled. I looked at the storge tags options but that involves adding tags to service offerings and I dont want to mess up with that tags. Should we add a new api to enable this feature? or any other better suggestion? -- Thanks and regards Rakesh
Re: Keyboard Shortcuts on Management Server.
I looked at the GSOC idea. I have feedback on this 1. Need to memorize the shortcut which is not so good 2. only a limited set of shortcuts as of now.. if it grows then more memorization If we use vimium or similar plugin cons need to spend time installing in browsers pros very less memorization more shortcuts no need for source code change if more tabs/panes or whatever is added to ui future. (just press F and you are good to go) On Wed, Jun 9, 2021 at 1:38 PM Boris Stoyanov wrote: > Not everyone uses vimium browser plugin, I think it would be great if we > can get this working cross-browser. > > Here’s the original GSOC Idea: > https://github.com/apache/cloudstack/issues/4798 > > Bobby. > > From: Rakesh Venkatesh > Date: Wednesday, 9 June 2021, 14:29 > To: dev > Subject: Re: Keyboard Shortcuts on Management Server. > I usually use vimium plugin which has keyboard shortcuts for all links in a > webpage which is much more handy > > > > > On Wed, Jun 9, 2021 at 12:38 PM Apurv Gupta wrote: > > > Hi all, > > > > > > > > I’m trying to implement the keyboard shortcuts and the idea is to use the > > management server ui without a mouse. > > > > From the mentors, here are the suggestion list for shortcuts: > > > > > > > > [i] - infrastructure > > > > [g] - configuration->global settings > > > > [v] - storage->volumes > > > > [c] - compute->instances > > > > [t] - images->templates > > > > [n] - networks > > > > [a] - accounts > > > > [p] - projects > > > > [e] - events > > > > > > > > Please feel free to add more to the list and would appreciate any > comments, > > suggestions or objections. > > Also, I’ve opened a draft PR #5090 ( > > https://github.com/apache/cloudstack/pull/5090). > > > > > > > > Thankyou! > > > > > -- > Thanks and regards > Rakesh venkatesh > -- Thanks and regards Rakesh venkatesh
Re: Limiting the conntrackd logs in virtual router
If I restart with cleanup then the modified config will be deleted and the log starts populating very fast again. On Wed, Jun 9, 2021 at 1:19 PM Andrija Panic wrote: > Did you try restarting that VPC with cleanup - so brand new VR is created - > to see if the issue is still observed. > > Can you test the same now with 4.15.1 RC1 (voting is happening, please feel > free to test and participate!) > > Best, > > On Wed, 9 Jun 2021 at 10:57, Rakesh Venkatesh > wrote: > > > I have 4.14 version. The issue exists with only 1 customer. > > The VR's belonging to other customers are fine but for one customer who > has > > enabled site 2 site vpn, its creating lot of logs. > > The config I mentioned above is a temporary fix as it will be deleted > once > > the network is restarted with cleanup. > > > > On Wed, Jun 9, 2021 at 10:52 AM Andrija Panic > > wrote: > > > > > Can you advise which version and have you tried isolating the problem - > > > this is first time I hear for such an issue (and I have customers with > 3y > > > VR uptime with no issues) > > > > > > Perhaps something is broken inside that VR, thus the excessive logging, > > > etc? > > > > > > Best, > > > > > > On Wed, 9 Jun 2021 at 10:07, Rakesh Venkatesh < > www.rakeshv@gmail.com > > > > > > wrote: > > > > > > > Hello > > > > > > > > I have an issue in the virtual router where > conntrackd(conntrack-tools) > > > is > > > > populating /var/log/daemon.log a lot and because of that VR is > running > > > out > > > > of disk space. Do you guys know how to prevent conntrackd from > logging > > > too > > > > much? I manually changed the logrotate.conf to rotate every hour and > > > retain > > > > only few copies but thats a quick fix. Is there any better fix to > avoid > > > > this issue? In our setup the VR has 2GB disk and even that got filled > > up > > > > pretty quickly. > > > > > > > > The logrotate config for daemon.log is present in > > > /etc/logrotate.d/rsyslog > > > > and I added > > > > > > > > /var/log/daemon.log > > > > { > > > > hourly > > > > rotate 1 > > > > missingok > > > > notifempty > > > > compress > > > > copytruncate > > > > } > > > > > > > > > > > > The config present in /etc/logrotate.d/conntrackd is pointing > > > > to /var/log/conntrackd-stats.log and so changing that config wont > help. > > > > > > > > -- > > > > Thanks and regards > > > > Rakesh > > > > > > > > > > > > > -- > > > > > > Andrija Panić > > > > > > > > > -- > > Thanks and regards > > Rakesh venkatesh > > > > > -- > > Andrija Panić > -- Thanks and regards Rakesh venkatesh
Re: Limiting the conntrackd logs in virtual router
I have 4.14 version. The issue exists with only 1 customer. The VR's belonging to other customers are fine but for one customer who has enabled site 2 site vpn, its creating lot of logs. The config I mentioned above is a temporary fix as it will be deleted once the network is restarted with cleanup. On Wed, Jun 9, 2021 at 10:52 AM Andrija Panic wrote: > Can you advise which version and have you tried isolating the problem - > this is first time I hear for such an issue (and I have customers with 3y > VR uptime with no issues) > > Perhaps something is broken inside that VR, thus the excessive logging, > etc? > > Best, > > On Wed, 9 Jun 2021 at 10:07, Rakesh Venkatesh > wrote: > > > Hello > > > > I have an issue in the virtual router where conntrackd(conntrack-tools) > is > > populating /var/log/daemon.log a lot and because of that VR is running > out > > of disk space. Do you guys know how to prevent conntrackd from logging > too > > much? I manually changed the logrotate.conf to rotate every hour and > retain > > only few copies but thats a quick fix. Is there any better fix to avoid > > this issue? In our setup the VR has 2GB disk and even that got filled up > > pretty quickly. > > > > The logrotate config for daemon.log is present in > /etc/logrotate.d/rsyslog > > and I added > > > > /var/log/daemon.log > > { > > hourly > > rotate 1 > > missingok > > notifempty > > compress > > copytruncate > > } > > > > > > The config present in /etc/logrotate.d/conntrackd is pointing > > to /var/log/conntrackd-stats.log and so changing that config wont help. > > > > -- > > Thanks and regards > > Rakesh > > > > > -- > > Andrija Panić > -- Thanks and regards Rakesh venkatesh
Limiting the conntrackd logs in virtual router
Hello I have an issue in the virtual router where conntrackd(conntrack-tools) is populating /var/log/daemon.log a lot and because of that VR is running out of disk space. Do you guys know how to prevent conntrackd from logging too much? I manually changed the logrotate.conf to rotate every hour and retain only few copies but thats a quick fix. Is there any better fix to avoid this issue? In our setup the VR has 2GB disk and even that got filled up pretty quickly. The logrotate config for daemon.log is present in /etc/logrotate.d/rsyslog and I added /var/log/daemon.log { hourly rotate 1 missingok notifempty compress copytruncate } The config present in /etc/logrotate.d/conntrackd is pointing to /var/log/conntrackd-stats.log and so changing that config wont help. -- Thanks and regards Rakesh
SSL handshake error while connecting to mgt server
Hello Host is failing to connect to mgt server with the following error message. Any idea what could be the reason. It started happening all of a sudden. I have other hosts in the platform and they are working fine. 2021-05-20 12:34:31,725 ERROR [utils.nio.NioConnection] (main:null) Unable to initialize the threads. javax.net.ssl.SSLHandshakeException: No appropriate protocol (protocol is disabled or cipher suites are inappropriate) > at sun.security.ssl.HandshakeContext.(HandshakeContext.java:171) > > at > sun.security.ssl.ClientHandshakeContext.(ClientHandshakeContext.java:98) > at sun.security.ssl.TransportContext.kickstart(TransportContext.java:220) > at sun.security.ssl.SSLEngineImpl.writeRecord(SSLEngineImpl.java:159) > at sun.security.ssl.SSLEngineImpl.wrap(SSLEngineImpl.java:130) > at sun.security.ssl.SSLEngineImpl.wrap(SSLEngineImpl.java:110) > at javax.net.ssl.SSLEngine.wrap(SSLEngine.java:471) > at com.cloud.utils.nio.Link.doHandshake(Link.java:489) > at com.cloud.utils.nio.NioClient.init(NioClient.java:80) > at com.cloud.utils.nio.NioConnection.start(NioConnection.java:88) > at com.cloud.agent.Agent.start(Agent.java:246) > at com.cloud.agent.AgentShell.launchAgent(AgentShell.java:448) > at com.cloud.agent.AgentShell.launchAgentFromClassInfo(AgentShell.java:416) > at com.cloud.agent.AgentShell.launchAgent(AgentShell.java:400) > at com.cloud.agent.AgentShell.start(AgentShell.java:510) > at com.cloud.agent.AgentShell.main(AgentShell.java:545) > > -- Thanks and regards Rakesh
Changing the role of the account
Hello folks I dont think there is an endpoint right now to change the role of the account. If I have to change that in db, that should be doable right by changing the id's in two or three tables. I hope that wont break anything else. Is it good to add a new parameter to updateAccount api to take the new role id? -- Thanks and regards Rakesh venkatesh
Re: Db schema upgrade fails from 4.14 to 4.16
Hello Pearl No, this was the first time I tried to upgrade from 4.14 to master. 4.14 was running for quite long time. I took db backup, registered 4.16 systemvm, changed global settings and then just deployed new packages. It failed in the first attempt. I was able to resolve this by dropping the project_account and project_invitation table, creating them and restarted the mgt service. This time it applied successfully To double check it, I created a fresh 4.14 system, did the above steps and the upgrade still failed. So solution right now is to drop the tables, create it again and run the upgrade On Thu, Apr 8, 2021 at 6:52 PM Pearl d'Silva wrote: > Hi Rakesh, > > Was there a prior attempt made to upgrade which didn't go through > successfully? It seems like while it failed to upgrade, the DB changes were > applied, hence resulting in the failure that is noticed. To address the > issue, you will have to restore the DB, if you've taken a backup of it and > redo the upgrade. > However, to avoid such situations, an enhancement was added with > https://github.com/apache/cloudstack/pull/4582 which prevents upgrading > the DB in case of missing systemvm template. > > Thanks > Pearl > > > From: Andrija Panic > Sent: Thursday, April 8, 2021 9:09 PM > To: dev > Cc: users > Subject: Re: Db schema upgrade fails from 4.14 to 4.16 > > 4.16 doesn't exist yet, and it's the master branch - so failures are > expected... > > On Thu, 8 Apr 2021 at 16:42, Rakesh Venkatesh http://www.rakeshv@gmail.com>> > wrote: > > > Hello users > > > > > > Im trying to upgrade cloudstack from 4.14.0 to 4.16.0 and the upgrade is > > failing because of the db schema upgrade. Any idea on how to fix this. > > > > This is the commit which is causing the issue > > https://github.com/apache/cloudstack/pull/4128 > > > > > > > > 2021-04-08 13:59:26,983 DEBUG [c.c.u.d.ScriptRunner] (main:null) (logid:) > > -- Schema upgrade from 4.14.0.0 to 4.15.0.0 > > > > 2021-04-08 13:59:27,002 ERROR [c.c.u.d.ScriptRunner] (main:null) (logid:) > > Error executing: ALTER TABLE `cloud`.`project_account` ADD COLUMN > > `user_id` bigint unsigned COMMENT 'ID of user to be added to the project' > > AFTER `account_id`, ADD CONSTRAINT `fk_project_account__user_id` FOREIGN > > KEY `fk_project_account__user_id`(`user_id`) REFERENCES `user`(`id`) ON > > DELETE CASCADE, ADD COLUMN `project_role_id` bigint unsigned COMMENT > > 'Project role id' AFTER `project_account_id`, ADD CONSTRAINT > > `fk_project_account__project_role_id` FOREIGN KEY (`project_role_id`) > > REFERENCES `project_role` (`id`) ON DELETE SET NULL, DROP FOREIGN KEY > > `fk_project_account__account_id`, DROP INDEX `account_id` > > 2021-04-08 13:59:27,003 ERROR [c.c.u.d.ScriptRunner] (main:null) (logid:) > > java.sql.SQLSyntaxErrorException: Can't DROP > > 'fk_project_account__account_id'; check that column/key exists > > 2021-04-08 13:59:27,004 ERROR [c.c.u.DatabaseUpgradeChecker] (main:null) > > (logid:) Unable to execute upgrade script > > java.sql.SQLSyntaxErrorException: Can't DROP > > 'fk_project_account__account_id'; check that column/key exists > > at com.cloud.utils.db.ScriptRunner.runScript(ScriptRunner.java:185) > > > > > > Caused by: java.sql.SQLSyntaxErrorException: Can't DROP > > 'fk_project_account__account_id'; check that column/key exists > > -- > > Thanks and regards > > Rakesh > > > > > -- > > Andrija Panić > > pearl.dsi...@shapeblue.com > www.shapeblue.com > 3 London Bridge Street, 3rd floor, News Building, London SE1 9SGUK > @shapeblue > > > > -- Thanks and regards Rakesh venkatesh
Db schema upgrade fails from 4.14 to 4.16
Hello users Im trying to upgrade cloudstack from 4.14.0 to 4.16.0 and the upgrade is failing because of the db schema upgrade. Any idea on how to fix this. This is the commit which is causing the issue https://github.com/apache/cloudstack/pull/4128 2021-04-08 13:59:26,983 DEBUG [c.c.u.d.ScriptRunner] (main:null) (logid:) -- Schema upgrade from 4.14.0.0 to 4.15.0.0 2021-04-08 13:59:27,002 ERROR [c.c.u.d.ScriptRunner] (main:null) (logid:) Error executing: ALTER TABLE `cloud`.`project_account` ADD COLUMN `user_id` bigint unsigned COMMENT 'ID of user to be added to the project' AFTER `account_id`, ADD CONSTRAINT `fk_project_account__user_id` FOREIGN KEY `fk_project_account__user_id`(`user_id`) REFERENCES `user`(`id`) ON DELETE CASCADE, ADD COLUMN `project_role_id` bigint unsigned COMMENT 'Project role id' AFTER `project_account_id`, ADD CONSTRAINT `fk_project_account__project_role_id` FOREIGN KEY (`project_role_id`) REFERENCES `project_role` (`id`) ON DELETE SET NULL, DROP FOREIGN KEY `fk_project_account__account_id`, DROP INDEX `account_id` 2021-04-08 13:59:27,003 ERROR [c.c.u.d.ScriptRunner] (main:null) (logid:) java.sql.SQLSyntaxErrorException: Can't DROP 'fk_project_account__account_id'; check that column/key exists 2021-04-08 13:59:27,004 ERROR [c.c.u.DatabaseUpgradeChecker] (main:null) (logid:) Unable to execute upgrade script java.sql.SQLSyntaxErrorException: Can't DROP 'fk_project_account__account_id'; check that column/key exists at com.cloud.utils.db.ScriptRunner.runScript(ScriptRunner.java:185) Caused by: java.sql.SQLSyntaxErrorException: Can't DROP 'fk_project_account__account_id'; check that column/key exists -- Thanks and regards Rakesh
Automatically applying DB schema changes
Hello Users and Dev Is there a way a new DB schema changes can be applied automatically whenever I install new packages? My setup was running with two month old changes of 4.15 and when I deployed new packages with latest changes, all the recent db scheme changes are not applied and I need to run it manually. How do I avoid it and how do you guys do it? For example: This is the error I get Caused by: java.sql.SQLSyntaxErrorException: Unknown column 'image_store.readonly' in 'field list' and a big stack trace This was fixed by applying changes from https://github.com/apache/cloudstack/blob/master/engine/schema/src/main/resources/META-INF/db/schema-41400to41500.sql#L198-L222 Another error was Caused by: java.sql.SQLSyntaxErrorException: Unknown column 'project_invitations.account_role' in 'field list' So I had to apply the schema needed for project_role related queries -- Thanks and regards Rakesh
Deploying Linux/Windows Vm's on hypervisors
Hello All I want your opinion on how you organize deploying different types of VM's(Linux, windows) on a specific set of hypervisors. In our setup, we have added a tag "Linux,Unix" on few hypervisors and also on templates so that when we create a vm using these templates then it will be deployed on HV's matching that tag only. If there is no tag on the template then it can be deployed on any hypervisor. We have some windows templates which do not have a tag as they are uploaded by customers and when VM is created from that template, it will be deployed in any suitable hypervisor but we want it to be deployed on a hypervisor having "Windows" tag. Is there any way to achieve this apart from using template tag? Another way is to use the OS type and match with hypervisor tags but that may not be a suitable solution. So if anyone has managed to deploy windows vm's on windows tagged hypervisor and linux vm's on linux/unix tagged hypervisor then please let me know -- Thanks and regards Rakesh
Re: Unable to ping secondary NIC ip of a VM in VPC
Hi Andrija Yes if I configure eth1 manually inside VM to get Ip from dhcp then ping works fine but not sure how to do this inside windows vm On Wed, Aug 12, 2020 at 10:43 PM Andrija Panic wrote: > Hi Rakesh, > > if you configure eth1 manually inside a VM, but set it to obtain the IP via > DHCP, does it obtain the IP properly then? > > Best, > > On Tue, 11 Aug 2020, 12:41 Rakesh Venkatesh, > wrote: > > > Wei, > > > > This was the test done on ubuntu os. The initial issue was found on > windows > > vm. > > > > Even if it can be solved by adding some config inside the vm, do we need > to > > do every time we delete/add additional NIC's to vm? That will be extra > work > > all the time right? > > > > On Mon, Aug 10, 2020 at 5:43 PM Wei ZHOU wrote: > > > > > Hi Rakesh, > > > > > > It is not a bug. you need to configure the networks inside vm. for > > example > > > /etc/network/interfaces in old ubuntu versions or /etc/netplan in newer > > > ubuntu versions. > > > > > > -Wei > > > > > > On Monday, 10 August 2020, Rakesh Venkatesh > > > > wrote: > > > > > > > Hello Thomas > > > > > > > > > > > > Thanks for the reply. It looks like the NIC is not configured > properly > > in > > > > VM. I dont see IP 10.10.2.23 being configured > > > > > > > > # ip a > > > > 1: lo: mtu 65536 qdisc noqueue state UNKNOWN > > group > > > > default qlen 1 > > > > link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 > > > > inet 127.0.0.1/8 scope host lo > > > >valid_lft forever preferred_lft forever > > > > inet6 ::1/128 scope host > > > >valid_lft forever preferred_lft forever > > > > 2: eth0: mtu 1500 qdisc pfifo_fast > > > state > > > > UP group default qlen 1000 > > > > link/ether 02:00:78:24:00:01 brd ff:ff:ff:ff:ff:ff > > > > inet 10.10.1.51/24 brd 10.10.1.255 scope global eth0 > > > >valid_lft forever preferred_lft forever > > > > inet6 fe80::78ff:fe24:1/64 scope link > > > >valid_lft forever preferred_lft forever > > > > 3: eth1: mtu 1500 qdisc noop state DOWN group > > > default > > > > qlen 1000 > > > > link/ether 02:00:1f:ed:00:0e brd ff:ff:ff:ff:ff:ff > > > > > > > > > > > > > > > > I removed the secondary NIC, stopped the vm, added secondary NIC and > > > > started the VM. The extra NIC is still not configured in VM. > > > > Is this a bug? > > > > > > > > On Mon, Aug 10, 2020 at 4:46 PM Thomas Joseph < > thomas.jo...@gmail.com> > > > > wrote: > > > > > > > > > Hello Venkatesh, > > > > > > > > > > I'm guessing from the output provided that you have tried to ping > the > > > ip > > > > > from the VR. > > > > > Have you checked if the NIC is up on the VM? > > > > > you can run tcpdump on the related VM to check if the icmp packets > > are > > > > > reaching the VM. > > > > > > > > > > Regards > > > > > Thomas > > > > > > > > > > On Mon, 10 Aug 2020, 8:06 pm Rakesh Venkatesh, < > > > > www.rakeshv@gmail.com> > > > > > wrote: > > > > > > > > > > > Hello, > > > > > > > > > > > > I have created a VPC with three tiers in it (tier1, tier2. tier3) > > > using > > > > > > cloudstack 4.14.0.0 version with Ubuntu16 as OS for mgt, agent > > > > > > I created a VM in tier1 and I am able to ping its IP address > > > > > > > > > > > > # ping 10.10.1.51 > > > > > > PING 10.10.1.51 (10.10.1.51): 56 data bytes > > > > > > 64 bytes from 10.10.1.51: icmp_seq=0 ttl=64 time=2.273 ms > > > > > > 64 bytes from 10.10.1.51: icmp_seq=1 ttl=64 time=1.690 ms > > > > > > ^C--- 10.10.1.51 ping statistics --- > > > > > > > > > > > > > > > > > > After adding secondary NIC to vm from tier2 or tier3 then I am > not > > > able > > > > > to > > > > > > ping that IP address. > > > > > > > > > > > > # ping 10.10.2.23 > > > &g
Re: How to delete OS template via database
vm_template is not an api. Its a table in cloud database On Wed, Aug 12, 2020 at 11:35 AM Wan Rahim, Wan Mohd Saifudin < wanmohdsaifudin.wanra...@itelligence.com.my> wrote: > Already tested to delete the template in cloudmonkey but still failed. > > And vm_template api is not available. > > Is there any other method can be used? > > Thanks > Din > > -Original Message- > From: Rakesh Venkatesh [mailto:www.rakeshv@gmail.com] > Sent: Tuesday, 11 August, 2020 6:47 PM > To: users > Subject: Re: How to delete OS template via database > > EXTERNAL MAIL: > > You can set "state" to "Inactive" in vm_template table and it wont show up > in UI > > update vm_template set state="Inactive" where id=<> limit 1; > > or you can run delete template id= cloudmonkey api to delete it > > On Tue, Aug 11, 2020 at 9:49 AM harith roli wrote: > > > please help, I need a way to delete OS template of cloudstack via > > database. Currently the status of template is UploadAbandoned. I try > > delete using gui, Zones --> click button "x" to delete template but I > > got message box indicate status failed to delete template. Thank you > > for your help > > > > > -- > Thanks and regards > Rakesh venkatesh > > > -- Thanks and regards Rakesh venkatesh
Re: How to delete OS template via database
You can set "state" to "Inactive" in vm_template table and it wont show up in UI update vm_template set state="Inactive" where id=<> limit 1; or you can run delete template id= cloudmonkey api to delete it On Tue, Aug 11, 2020 at 9:49 AM harith roli wrote: > please help, I need a way to delete OS template of cloudstack via > database. Currently the status of template is UploadAbandoned. I try delete > using gui, Zones --> click button "x" to delete template but I got message > box indicate status failed to delete template. Thank you for your help > -- Thanks and regards Rakesh venkatesh
Re: Unable to ping secondary NIC ip of a VM in VPC
Wei, This was the test done on ubuntu os. The initial issue was found on windows vm. Even if it can be solved by adding some config inside the vm, do we need to do every time we delete/add additional NIC's to vm? That will be extra work all the time right? On Mon, Aug 10, 2020 at 5:43 PM Wei ZHOU wrote: > Hi Rakesh, > > It is not a bug. you need to configure the networks inside vm. for example > /etc/network/interfaces in old ubuntu versions or /etc/netplan in newer > ubuntu versions. > > -Wei > > On Monday, 10 August 2020, Rakesh Venkatesh > wrote: > > > Hello Thomas > > > > > > Thanks for the reply. It looks like the NIC is not configured properly in > > VM. I dont see IP 10.10.2.23 being configured > > > > # ip a > > 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group > > default qlen 1 > > link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 > > inet 127.0.0.1/8 scope host lo > >valid_lft forever preferred_lft forever > > inet6 ::1/128 scope host > >valid_lft forever preferred_lft forever > > 2: eth0: mtu 1500 qdisc pfifo_fast > state > > UP group default qlen 1000 > > link/ether 02:00:78:24:00:01 brd ff:ff:ff:ff:ff:ff > > inet 10.10.1.51/24 brd 10.10.1.255 scope global eth0 > >valid_lft forever preferred_lft forever > > inet6 fe80::78ff:fe24:1/64 scope link > >valid_lft forever preferred_lft forever > > 3: eth1: mtu 1500 qdisc noop state DOWN group > default > > qlen 1000 > > link/ether 02:00:1f:ed:00:0e brd ff:ff:ff:ff:ff:ff > > > > > > > > I removed the secondary NIC, stopped the vm, added secondary NIC and > > started the VM. The extra NIC is still not configured in VM. > > Is this a bug? > > > > On Mon, Aug 10, 2020 at 4:46 PM Thomas Joseph > > wrote: > > > > > Hello Venkatesh, > > > > > > I'm guessing from the output provided that you have tried to ping the > ip > > > from the VR. > > > Have you checked if the NIC is up on the VM? > > > you can run tcpdump on the related VM to check if the icmp packets are > > > reaching the VM. > > > > > > Regards > > > Thomas > > > > > > On Mon, 10 Aug 2020, 8:06 pm Rakesh Venkatesh, < > > www.rakeshv@gmail.com> > > > wrote: > > > > > > > Hello, > > > > > > > > I have created a VPC with three tiers in it (tier1, tier2. tier3) > using > > > > cloudstack 4.14.0.0 version with Ubuntu16 as OS for mgt, agent > > > > I created a VM in tier1 and I am able to ping its IP address > > > > > > > > # ping 10.10.1.51 > > > > PING 10.10.1.51 (10.10.1.51): 56 data bytes > > > > 64 bytes from 10.10.1.51: icmp_seq=0 ttl=64 time=2.273 ms > > > > 64 bytes from 10.10.1.51: icmp_seq=1 ttl=64 time=1.690 ms > > > > ^C--- 10.10.1.51 ping statistics --- > > > > > > > > > > > > After adding secondary NIC to vm from tier2 or tier3 then I am not > able > > > to > > > > ping that IP address. > > > > > > > > # ping 10.10.2.23 > > > > PING 10.10.2.23 (10.10.2.23): 56 data bytes > > > > 92 bytes from 10.10.2.254: Destination Host Unreachable > > > > 92 bytes from 10.10.2.254: Destination Host Unreachable > > > > 92 bytes from 10.10.2.254: Destination Host Unreachable > > > > 92 bytes from 10.10.2.254: Destination Host Unreachable > > > > ^C--- 10.10.2.23 ping statistics --- > > > > > > > > > > > > > > > > Dchphost content > > > > > > > > # cat /etc/dhcphosts.txt > > > > 02:00:72:36:00:01,10.10.2.83,tier12,infinite > > > > 02:00:2f:16:00:0c,set:10_10_2_23,10.10.2.23,tier11,infinite > > > > 02:00:78:24:00:01,10.10.1.51,tier11,infinite > > > > > > > > > > > > > > > > Dhcpopts content > > > > > > > > > > > > # cat /etc/dhcpopts.txt > > > > 10_10_2_23,3 > > > > 10_10_2_23,6 > > > > 10_10_2_23,15 > > > > > > > > > > > > I don't see any errors in /var/log/cloud.log on VR when secondary nic > > is > > > > added. > > > > Is there any other I can troubleshoot this or look for logs? > > > > > > > > -- > > > > Thanks and regards > > > > Rakesh > > > > > > > > > > > > > -- > > Thanks and regards > > Rakesh venkatesh > > > -- Thanks and regards Rakesh venkatesh
Re: Unable to ping secondary NIC ip of a VM in VPC
Ok no problem. As suggested, I manually edited /etc/network/interfaces files to add eth1 and restarted networking service. Ping works fine now from VR # ping 10.10.2.25 PING 10.10.2.25 (10.10.2.25): 56 data bytes 64 bytes from 10.10.2.25: icmp_seq=0 ttl=64 time=1.981 ms 64 bytes from 10.10.2.25: icmp_seq=1 ttl=64 time=1.322 ms 64 bytes from 10.10.2.25: icmp_seq=2 ttl=64 time=1.124 ms 64 bytes from 10.10.2.25: icmp_seq=3 ttl=64 time=1.376 ms ^C--- 10.10.2.25 ping statistics --- 4 packets transmitted, 4 packets received, 0% packet loss On Mon, Aug 10, 2020 at 5:04 PM Thomas Joseph wrote: > Sorry, I've not used v4.14, so someone else can comment on it being a bug. > > I would advise you to enable the related nic in the file named interfaces > and restart the network service to check if that works. > > > On Mon, 10 Aug 2020, 8:29 pm Rakesh Venkatesh, > wrote: > > > Hello Thomas > > > > > > Thanks for the reply. It looks like the NIC is not configured properly in > > VM. I dont see IP 10.10.2.23 being configured > > > > # ip a > > 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group > > default qlen 1 > > link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 > > inet 127.0.0.1/8 scope host lo > >valid_lft forever preferred_lft forever > > inet6 ::1/128 scope host > >valid_lft forever preferred_lft forever > > 2: eth0: mtu 1500 qdisc pfifo_fast > state > > UP group default qlen 1000 > > link/ether 02:00:78:24:00:01 brd ff:ff:ff:ff:ff:ff > > inet 10.10.1.51/24 brd 10.10.1.255 scope global eth0 > >valid_lft forever preferred_lft forever > > inet6 fe80::78ff:fe24:1/64 scope link > >valid_lft forever preferred_lft forever > > 3: eth1: mtu 1500 qdisc noop state DOWN group > default > > qlen 1000 > > link/ether 02:00:1f:ed:00:0e brd ff:ff:ff:ff:ff:ff > > > > > > > > I removed the secondary NIC, stopped the vm, added secondary NIC and > > started the VM. The extra NIC is still not configured in VM. > > Is this a bug? > > > > On Mon, Aug 10, 2020 at 4:46 PM Thomas Joseph > > wrote: > > > > > Hello Venkatesh, > > > > > > I'm guessing from the output provided that you have tried to ping the > ip > > > from the VR. > > > Have you checked if the NIC is up on the VM? > > > you can run tcpdump on the related VM to check if the icmp packets are > > > reaching the VM. > > > > > > Regards > > > Thomas > > > > > > On Mon, 10 Aug 2020, 8:06 pm Rakesh Venkatesh, < > > www.rakeshv@gmail.com> > > > wrote: > > > > > > > Hello, > > > > > > > > I have created a VPC with three tiers in it (tier1, tier2. tier3) > using > > > > cloudstack 4.14.0.0 version with Ubuntu16 as OS for mgt, agent > > > > I created a VM in tier1 and I am able to ping its IP address > > > > > > > > # ping 10.10.1.51 > > > > PING 10.10.1.51 (10.10.1.51): 56 data bytes > > > > 64 bytes from 10.10.1.51: icmp_seq=0 ttl=64 time=2.273 ms > > > > 64 bytes from 10.10.1.51: icmp_seq=1 ttl=64 time=1.690 ms > > > > ^C--- 10.10.1.51 ping statistics --- > > > > > > > > > > > > After adding secondary NIC to vm from tier2 or tier3 then I am not > able > > > to > > > > ping that IP address. > > > > > > > > # ping 10.10.2.23 > > > > PING 10.10.2.23 (10.10.2.23): 56 data bytes > > > > 92 bytes from 10.10.2.254: Destination Host Unreachable > > > > 92 bytes from 10.10.2.254: Destination Host Unreachable > > > > 92 bytes from 10.10.2.254: Destination Host Unreachable > > > > 92 bytes from 10.10.2.254: Destination Host Unreachable > > > > ^C--- 10.10.2.23 ping statistics --- > > > > > > > > > > > > > > > > Dchphost content > > > > > > > > # cat /etc/dhcphosts.txt > > > > 02:00:72:36:00:01,10.10.2.83,tier12,infinite > > > > 02:00:2f:16:00:0c,set:10_10_2_23,10.10.2.23,tier11,infinite > > > > 02:00:78:24:00:01,10.10.1.51,tier11,infinite > > > > > > > > > > > > > > > > Dhcpopts content > > > > > > > > > > > > # cat /etc/dhcpopts.txt > > > > 10_10_2_23,3 > > > > 10_10_2_23,6 > > > > 10_10_2_23,15 > > > > > > > > > > > > I don't see any errors in /var/log/cloud.log on VR when secondary nic > > is > > > > added. > > > > Is there any other I can troubleshoot this or look for logs? > > > > > > > > -- > > > > Thanks and regards > > > > Rakesh > > > > > > > > > > > > > -- > > Thanks and regards > > Rakesh venkatesh > > > -- Thanks and regards Rakesh venkatesh
Re: Unable to ping secondary NIC ip of a VM in VPC
Hello Thomas Thanks for the reply. It looks like the NIC is not configured properly in VM. I dont see IP 10.10.2.23 being configured # ip a 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 2: eth0: mtu 1500 qdisc pfifo_fast state UP group default qlen 1000 link/ether 02:00:78:24:00:01 brd ff:ff:ff:ff:ff:ff inet 10.10.1.51/24 brd 10.10.1.255 scope global eth0 valid_lft forever preferred_lft forever inet6 fe80::78ff:fe24:1/64 scope link valid_lft forever preferred_lft forever 3: eth1: mtu 1500 qdisc noop state DOWN group default qlen 1000 link/ether 02:00:1f:ed:00:0e brd ff:ff:ff:ff:ff:ff I removed the secondary NIC, stopped the vm, added secondary NIC and started the VM. The extra NIC is still not configured in VM. Is this a bug? On Mon, Aug 10, 2020 at 4:46 PM Thomas Joseph wrote: > Hello Venkatesh, > > I'm guessing from the output provided that you have tried to ping the ip > from the VR. > Have you checked if the NIC is up on the VM? > you can run tcpdump on the related VM to check if the icmp packets are > reaching the VM. > > Regards > Thomas > > On Mon, 10 Aug 2020, 8:06 pm Rakesh Venkatesh, > wrote: > > > Hello, > > > > I have created a VPC with three tiers in it (tier1, tier2. tier3) using > > cloudstack 4.14.0.0 version with Ubuntu16 as OS for mgt, agent > > I created a VM in tier1 and I am able to ping its IP address > > > > # ping 10.10.1.51 > > PING 10.10.1.51 (10.10.1.51): 56 data bytes > > 64 bytes from 10.10.1.51: icmp_seq=0 ttl=64 time=2.273 ms > > 64 bytes from 10.10.1.51: icmp_seq=1 ttl=64 time=1.690 ms > > ^C--- 10.10.1.51 ping statistics --- > > > > > > After adding secondary NIC to vm from tier2 or tier3 then I am not able > to > > ping that IP address. > > > > # ping 10.10.2.23 > > PING 10.10.2.23 (10.10.2.23): 56 data bytes > > 92 bytes from 10.10.2.254: Destination Host Unreachable > > 92 bytes from 10.10.2.254: Destination Host Unreachable > > 92 bytes from 10.10.2.254: Destination Host Unreachable > > 92 bytes from 10.10.2.254: Destination Host Unreachable > > ^C--- 10.10.2.23 ping statistics --- > > > > > > > > Dchphost content > > > > # cat /etc/dhcphosts.txt > > 02:00:72:36:00:01,10.10.2.83,tier12,infinite > > 02:00:2f:16:00:0c,set:10_10_2_23,10.10.2.23,tier11,infinite > > 02:00:78:24:00:01,10.10.1.51,tier11,infinite > > > > > > > > Dhcpopts content > > > > > > # cat /etc/dhcpopts.txt > > 10_10_2_23,3 > > 10_10_2_23,6 > > 10_10_2_23,15 > > > > > > I don't see any errors in /var/log/cloud.log on VR when secondary nic is > > added. > > Is there any other I can troubleshoot this or look for logs? > > > > -- > > Thanks and regards > > Rakesh > > > -- Thanks and regards Rakesh venkatesh
Unable to ping secondary NIC ip of a VM in VPC
Hello, I have created a VPC with three tiers in it (tier1, tier2. tier3) using cloudstack 4.14.0.0 version with Ubuntu16 as OS for mgt, agent I created a VM in tier1 and I am able to ping its IP address # ping 10.10.1.51 PING 10.10.1.51 (10.10.1.51): 56 data bytes 64 bytes from 10.10.1.51: icmp_seq=0 ttl=64 time=2.273 ms 64 bytes from 10.10.1.51: icmp_seq=1 ttl=64 time=1.690 ms ^C--- 10.10.1.51 ping statistics --- After adding secondary NIC to vm from tier2 or tier3 then I am not able to ping that IP address. # ping 10.10.2.23 PING 10.10.2.23 (10.10.2.23): 56 data bytes 92 bytes from 10.10.2.254: Destination Host Unreachable 92 bytes from 10.10.2.254: Destination Host Unreachable 92 bytes from 10.10.2.254: Destination Host Unreachable 92 bytes from 10.10.2.254: Destination Host Unreachable ^C--- 10.10.2.23 ping statistics --- Dchphost content # cat /etc/dhcphosts.txt 02:00:72:36:00:01,10.10.2.83,tier12,infinite 02:00:2f:16:00:0c,set:10_10_2_23,10.10.2.23,tier11,infinite 02:00:78:24:00:01,10.10.1.51,tier11,infinite Dhcpopts content # cat /etc/dhcpopts.txt 10_10_2_23,3 10_10_2_23,6 10_10_2_23,15 I don't see any errors in /var/log/cloud.log on VR when secondary nic is added. Is there any other I can troubleshoot this or look for logs? -- Thanks and regards Rakesh
Re: Temporarily disabling API requests
Thanks Andrija. Yes that's the better way On Thu, Jan 9, 2020 at 10:13 AM Andrija Panic wrote: > You can try editing the User role (ACS 4.9+) and explicitly add the deny > "ACL" for those API calls which you want to deny > Later just remove the "ACL" rules. > > On Thu, 9 Jan 2020 at 09:59, Rakesh Venkatesh > wrote: > > > Hello folks > > > > > > Is there a better way to disable all the VM related API's for a short > > duration of time? Like I want to disable stop/start/migrate/destroy VM > API. > > I'm thinking of disabling the user so that the user cant send any API > > requests. After some time I will enable the user back. Is there any other > > better way to do it? > > > > -- > > Thanks and regards > > Rakesh > > > > > -- > > Andrija Panić > -- Thanks and regards Rakesh venkatesh
Temporarily disabling API requests
Hello folks Is there a better way to disable all the VM related API's for a short duration of time? Like I want to disable stop/start/migrate/destroy VM API. I'm thinking of disabling the user so that the user cant send any API requests. After some time I will enable the user back. Is there any other better way to do it? -- Thanks and regards Rakesh
Number of hosts in a cluster
Hello In the documentation, I see that there is a configuration requirement of not adding more than 16 hosts in a cluster: "For KVM, do not put more than 16 hosts in a cluster." In our setup, we have more than 18 nodes in a cluster without any issues so far. So I wanted to know what is the actual limit of hosts in a cluster. I didn't see any such value in global setting as well. If there is no such limitation then we need to update the documentation. -- Thanks and regards Rakesh venkatesh
Re: Virtual machines volume lock manager
Hi Andrija Sorry for the late reply. Im using 4.7 version of ACS. Qemu version 1:2.5+dfsg-5ubuntu10.40 Im not sure if ACS job failed or libvirt job as I didnt see into logs. Yes the vm will be in paused state during migration but after the failed migration, the same vm was in "running" state on two different hypervisors. We wrote a script to find out how duplicated vm's are running and found out that more than 5 vm's had this issue. On Mon, Oct 28, 2019 at 2:42 PM Andrija Panic wrote: > I've been running KVM public cloud up to recently and have never seen such > behaviour. > > What versions (ACS, qemu, libvrit) are you running? > > How does the migration fail - ACS job - or libvirt job? > destination VM is by default always in PAUSED state, until the migration is > finished - only then the destination VM (on the new host) will get RUNNING, > while previously pausing the original VM (on the old host). > > i,e. > phase1 source vm RUNNING, destination vm PAUSED (RAM content being > copied over... takes time...) > phase2 source vm PAUSED, destination vm PAUSED (last bits of RAM > content are migrated) > phase3 source vm destroyed, destination VM RUNNING. > > Andrija > > On Mon, 28 Oct 2019 at 14:26, Rakesh Venkatesh > wrote: > > > Hello Users > > > > > > Recently we have seen cases where when the Vm migration fails, cloudstack > > ends up running two instances of the same VM on different hypervisors. > The > > state will be "running" and not any other transition state. This will of > > course lead to corruption of disk. Does CloudStack has any option of > volume > > locking so that two instances of the same VM wont be running? > > Anyone else has faced this issue and found some solution to fix it? > > > > We are thinking of using "virtlockd" of libvirt or implementing custom > lock > > mechanisms. There are some pros and cons of the both the solutions and i > > want your feedback before proceeding further. > > > > -- > > Thanks and regards > > Rakesh venkatesh > > > > > -- > > Andrija Panić > -- Thanks and regards Rakesh venkatesh
Virtual machines volume lock manager
Hello Users Recently we have seen cases where when the Vm migration fails, cloudstack ends up running two instances of the same VM on different hypervisors. The state will be "running" and not any other transition state. This will of course lead to corruption of disk. Does CloudStack has any option of volume locking so that two instances of the same VM wont be running? Anyone else has faced this issue and found some solution to fix it? We are thinking of using "virtlockd" of libvirt or implementing custom lock mechanisms. There are some pros and cons of the both the solutions and i want your feedback before proceeding further. -- Thanks and regards Rakesh venkatesh
Re: Dedicated hosts for Domain/Account
Thanks for the quick reply. I was browsing through the code and found the following // check affinity group of type Explicit dedication exists. If No put // dedicated pod/cluster/host in avoid list List vmGroupMappings = _affinityGroupVMMapDao.findByVmIdType(vm.getId(), "ExplicitDedication"); if (vmGroupMappings != null && !vmGroupMappings.isEmpty()) { isExplicit = true; } So this feature will work only if vm's are associated with affinity groups. I created two vm's with same affinity group and after enabling the maintenance mode they were migrated to the other dedicated machines. So no need to create a github issue I guess. On Mon, Aug 12, 2019 at 5:04 PM Andrija Panic wrote: > Considering that manual VM LIVE migrations via CloudStack from > non-dedicated to dedicated SHOULD/DOES work - then I would say this is an > "unhandled" case, which indeed should be handled and live migration should > happen instead of stopping the VMs. > > I assume someone else might jump in - but if not, please raise GitHub > issues as a bug report. > > > Thx > > On Mon, 12 Aug 2019 at 16:52, Rakesh Venkatesh > wrote: > > > Hello > > > > In my cloudstack setup, I have three KVM hypervisors out of which two > > hypervisors are dedicated to Root/admin account and the third is not > > dedicated. When I enable the maintenance mode on the dedicated > hypervisor, > > it will always migrate the vm's from dedicated to non dedicated > hypervisor > > but not to second dedicated hypervisor. I dont think this is the expected > > behavior. Can any one please verify? The dedicated hypervisors will be > > added to avoid set and the deployment planning manager skips these > > hypervisors. > > > > If I dedicate the third hypervisor to different domain and enable the > > maintenance mode on the first hypervisor then all the vm's will be > stopped > > instead of migrating to second dedicated hypervisor of the same > > domain/account. > > > > > > I have highlighted the necessary logs in red. You can see from the logs > > that host with id 17 and 20 are dedicated but not 26. When maintenance > mode > > is enabled on host id 20, it skips 17 and 20 and migrates vm's to host id > > 26 > > > > > > > > 2019-08-12 14:35:23,754 DEBUG [c.c.d.DeploymentPlanningManagerImpl] > > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c) > > (logid:a16d7711) Deploy avoids pods: null, clusters: null, hosts: [20], > > pools: null > > 2019-08-12 14:35:23,757 DEBUG [c.c.d.DeploymentPlanningManagerImpl] > > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c) > > (logid:a16d7711) DeploymentPlanner allocation algorithm: > > com.cloud.deploy.FirstFitPlanner@6fecace4 > > 2019-08-12 14:35:23,757 DEBUG [c.c.d.DeploymentPlanningManagerImpl] > > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c) > > (logid:a16d7711) Trying to allocate a host and storage pools from dc:8, > > pod:8,cluster:null, requested cpu: 16000, requested ram: 8589934592 > > 2019-08-12 14:35:23,757 DEBUG [c.c.d.DeploymentPlanningManagerImpl] > > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c) > > (logid:a16d7711) Is ROOT volume READY (pool already allocated)?: Yes > > 2019-08-12 14:35:23,757 DEBUG [c.c.d.DeploymentPlanningManagerImpl] > > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c) > > (logid:a16d7711) This VM has last host_id specified, trying to choose the > > same host: 20 > > 2019-08-12 14:35:23,759 DEBUG [c.c.d.DeploymentPlanningManagerImpl] > > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c) > > (logid:a16d7711) The last host of this VM is in avoid set > > 2019-08-12 14:35:23,759 DEBUG [c.c.d.DeploymentPlanningManagerImpl] > > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c) > > (logid:a16d7711) Cannot choose the last host to deploy this VM > > 2019-08-12 14:35:23,759 DEBUG [c.c.d.FirstFitPlanner] > > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c) > > (logid:a16d7711) Searching resources only under specified Pod: 8 > > 2019-08-12 14:35:23,759 DEBUG [c.c.d.FirstFitPlanner] > > (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c) > > (logid:a16d7711) Listing clusters in order of aggregate capacity, that > have > > (atleast one host with) enough CPU and RAM capacity under this Pod: 8 > > 2019-08-12 14:35:23,761 DEBUG [c.c.d.DeploymentPlanningManagerImpl] > > (Work-Job-Executor-7:ctx-9f4363d1 job-473/job-246899 ctx-cef9b496) > >
Dedicated hosts for Domain/Account
-246119/job-246902 ctx-9dbb7241) (logid:b7e8e3a2) Is ROOT volume READY (pool already allocated)?: Yes 2019-08-12 14:35:23,766 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-8:ctx-1cc07ab1 job-246119/job-246902 ctx-9dbb7241) (logid:b7e8e3a2) This VM has last host_id specified, trying to choose the same host: 26 2019-08-12 14:35:23,780 DEBUG [c.c.d.DeploymentPlanningManagerImpl] (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c) (logid:a16d7711) Checking resources in Cluster: 8 under Pod: 8 2019-08-12 14:35:23,782 DEBUG [c.c.c.CapacityManagerImpl] (Work-Job-Executor-7:ctx-9f4363d1 job-473/job-246899 ctx-cef9b496) (logid:bbb870bf) Host: 26 has cpu capability (cpu:48, speed:2900) to support requested CPU: 1 and requested speed: 500 2019-08-12 14:35:23,782 DEBUG [c.c.c.CapacityManagerImpl] (Work-Job-Executor-7:ctx-9f4363d1 job-473/job-246899 ctx-cef9b496) (logid:bbb870bf) Checking if host: 26 has enough capacity for requested CPU: 500 and requested RAM: 536870912 , cpuOverprovisioningFactor: 1.0 2019-08-12 14:35:23,782 DEBUG [c.c.a.m.a.i.FirstFitAllocator] (Work-Job-Executor-9:ctx-786e4f7a job-246740/job-246905 ctx-73b6368c FirstFitRoutingAllocator) (logid:a16d7711) Looking for hosts in dc: 8 pod:8 cluster:8 -- Thanks and regards Rakesh venkatesh
Querying async job result
Hello I want to know what is the best way to query the async job result using queryAsyncJobResult api. According to the documentation in http://docs.cloudstack.apache.org/projects/archived-cloudstack-getting-started/en/latest/dev.html , the "jobstatus" of 1 means the command completed but im facing an issue where even though the command is still running, the "jobstatus" is always 1. Im running "prepareHostForMaintenance" command which returns the jobid. When I run queryAsyncJobResult for this jobid, the jobstatus will always be 1 even though the hypervisor is still not in maintenance mode. So can anyone tell me what is the best way to check if the hypervisor is in maintenance mode or not? Im using 4.11 version Below are the result which I get "resourcestate": "PrepareForMaintenance", "jobresultcode": 0, "jobresulttype": "object", "jobstatus": 1, -- Thanks and regards Rakesh venkatesh
Secondary Storage VM timeout issue every hour
to Host 183 timed out after 3600 2019-07-25 04:01:22,785 DEBUG [c.c.a.m.AgentManagerImpl] (AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to listener: com.cloud.network.router.VpcVirtualNetworkApplianceManagerImpl 2019-07-25 04:01:22,785 DEBUG [c.c.a.m.AgentManagerImpl] (AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to listener: com.cloud.storage.download.DownloadListener 2019-07-25 04:01:22,785 ERROR [c.c.s.StatsCollector] (StatsCollector-2:ctx-b55657a9) (logid:dafc4881) Error trying to retrieve storage stats com.cloud.utils.exception.CloudRuntimeException: Failed to send command, due to Agent:183, com.cloud.exception.OperationTimedoutException: Commands 7541559051008607242 to Host 183 timed out after 3600 at org.apache.cloudstack.storage.RemoteHostEndPoint.sendMessage(RemoteHostEndPoint.java:133) at com.cloud.server.StatsCollector$StorageCollector.runInContext(StatsCollector.java:1139) at org.apache.cloudstack.managed.context.ManagedContextRunnable$1.run(ManagedContextRunnable.java:49) at org.apache.cloudstack.managed.context.impl.DefaultManagedContext$1.call(DefaultManagedContext.java:56) at org.apache.cloudstack.managed.context.impl.DefaultManagedContext.callWithContext(DefaultManagedContext.java:103) at org.apache.cloudstack.managed.context.impl.DefaultManagedContext.runWithContext(DefaultManagedContext.java:53) at org.apache.cloudstack.managed.context.ManagedContextRunnable.run(ManagedContextRunnable.java:46) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 2019-07-25 04:01:22,786 DEBUG [c.c.a.m.AgentManagerImpl] (AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to listener: com.cloud.consoleproxy.ConsoleProxyListener 2019-07-25 04:01:22,789 DEBUG [c.c.a.m.AgentManagerImpl] (AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to listener: com.cloud.storage.LocalStoragePoolListener 2019-07-25 04:01:22,789 DEBUG [c.c.a.m.AgentManagerImpl] (AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to listener: com.cloud.storage.upload.UploadListener 2019-07-25 04:01:22,790 DEBUG [c.c.a.m.AgentManagerImpl] (AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to listener: com.cloud.capacity.StorageCapacityListener 2019-07-25 04:01:22,790 DEBUG [c.c.a.m.AgentManagerImpl] (AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to listener: com.cloud.capacity.ComputeCapacityListener 2019-07-25 04:01:22,790 DEBUG [c.c.a.m.AgentManagerImpl] (AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to listener: com.cloud.network.SshKeysDistriMonitor 2019-07-25 04:01:22,791 DEBUG [c.c.a.m.AgentManagerImpl] (AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to listener: com.cloud.network.router.VirtualNetworkApplianceManagerImpl 2019-07-25 04:01:22,791 DEBUG [c.c.a.m.AgentManagerImpl] (AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to listener: com.cloud.network.NetworkUsageManagerImpl$DirectNetworkStatsListener 2019-07-25 04:01:22,791 DEBUG [c.c.n.NetworkUsageManagerImpl] (AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Disconnected called on 183 with status Disconnected 2019-07-25 04:01:22,791 DEBUG [c.c.a.m.AgentManagerImpl] (AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to listener: com.cloud.agent.manager.AgentManagerImpl$BehindOnPingListener 2019-07-25 04:01:22,791 DEBUG [c.c.a.m.AgentManagerImpl] (AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Sending Disconnect to listener: com.cloud.agent.manager.AgentManagerImpl$SetHostParamsListener 2019-07-25 04:01:22,791 DEBUG [c.c.h.Status] (AgentTaskPool-1:ctx-66de2057) (logid:841d2a63) Transition:[Resource state = Enabled, Agent event = ShutdownRequested, Host id = 183, name = s-2775-VM] -- Thanks and regards Rakesh venkatesh
Get VM OS type
Hello Folks Is there a way to know whether the VM is running on Windows or Linux OS? I can't reply on OS type because we can use Ubuntu as OS type for Windows VM. Even though the os type is Linux/Ubuntu, the VM is running on Windows version. Another way is to open the console and see the login screen. This will get the actual data but I want to do automation to see for all VM's and opening the console is not feasible to automate. Is there any other way to get it? -- Thanks and regards Rakesh Venkatesh
Switching backup VR to master manually
Hello Is there a way to switch the backup virtual router to master state manually by triggering something or running some scripts? Currently when I restart network with cleanup option, first the backup will be upgraded to newer cloudstack version and when the master is destroyed then only backup becomes the master and this process might take atleast 5-10 seconds. In the meantime there will be dowmtime. Is there a way to quickly change the state? Im thinking on stopping the keeaplived on master so that backup will become master quickly. Is there a better way to do it? -- Thanks and regards Rakesh venkatesh
Re: Troubleshooting the Exception "Unable to start VM instance"
Hello Peter If you dig into logs a little bit more, you can see the actual reason for the error. The usual errors are insufficient server capacity exception like not enough cpu, memory, ip address or mismatch between template tag and host tag or some other mismatch in the tags. These are the common errors I have seen. Search for "startVirtualMachine" api and trace its jobid which is usually like "context-" and look at the logs On Mon, Apr 29, 2019 at 12:58 PM wrote: > Hi, all, > > sometimes the system can't start a VM instance (a 530 error). > In the logs there is the following trace - how to troubleshoot it and > understand which circumstances lead to it? > > > com.cloud.utils.exception.CloudRuntimeException: Unable to start VM > instance > at > com.cloud.vm.VirtualMachineManagerImpl.orchestrateStart(VirtualMachineManagerImpl.java:4889) > at sun.reflect.GeneratedMethodAccessor433.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > com.cloud.vm.VmWorkJobHandlerProxy.handleVmWorkJob(VmWorkJobHandlerProxy.java:107) > at > com.cloud.vm.VirtualMachineManagerImpl.handleVmWorkJob(VirtualMachineManagerImpl.java:5047) > at > com.cloud.vm.VmWorkJobDispatcher.runJob(VmWorkJobDispatcher.java:102) > at > org.apache.cloudstack.framework.jobs.impl.AsyncJobManagerImpl$5.runInContext(AsyncJobManagerImpl.java:615) > at > org.apache.cloudstack.managed.context.ManagedContextRunnable$1.run(ManagedContextRunnable.java:49) > at > org.apache.cloudstack.managed.context.impl.DefaultManagedContext$1.call(DefaultManagedContext.java:56) > at > org.apache.cloudstack.managed.context.impl.DefaultManagedContext.callWithContext(DefaultManagedContext.java:103) > at > org.apache.cloudstack.managed.context.impl.DefaultManagedContext.runWithContext(DefaultManagedContext.java:53) > at > org.apache.cloudstack.managed.context.ManagedContextRunnable.run(ManagedContextRunnable.java:46) > at > org.apache.cloudstack.framework.jobs.impl.AsyncJobManagerImpl$5.run(AsyncJobManagerImpl.java:563) > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > at java.lang.Thread.run(Thread.java:748) > > kind regards > Peter > > -- Thanks and regards Rakesh venkatesh
Re: Cloudstack database timeout value
Thanks Suresh. I will try that out On Wed, Apr 10, 2019 at 6:32 PM Suresh Kumar Anaparti < sureshkumar.anapa...@gmail.com> wrote: > Hi Rakesh, > > Locate "db.properties" file (may be at /etc/cloudstack/management/) in the > management server and check for the parameters db.cloud.maxWait (default: > 10 secs), db.cloud.autoReconnect (default: true) and other database > settings. Restart the management server post parameters update. > > - Suresh > > On Wed, Apr 10, 2019 at 7:41 PM Rakesh Venkatesh < > www.rakeshv@gmail.com> > wrote: > > > Hello Folks > > > > Is there a way to increase the database connectivity timeout value in > case > > cloudstack loses connection to the database? Also is it possible to > > configure the value? Can you point me to the right place? I also want to > > increase the retries value so that it will retry to connect to DB > > configured value of times and then really shuts down. > > > > We have a redundant setup with three database hosts and switching from > one > > db host to another db host took more than 10 seconds and mgt server got > > killed since it cant connect to db anymore. i have to restart the mgt > > server to bring it back up. To avoid this I want to configure the timeout > > value as well as the number of retries to get a connection > > > > -- > > Thanks and regards > > Rakesh venkatesh > > > -- Thanks and regards Rakesh venkatesh
Cloudstack database timeout value
Hello Folks Is there a way to increase the database connectivity timeout value in case cloudstack loses connection to the database? Also is it possible to configure the value? Can you point me to the right place? I also want to increase the retries value so that it will retry to connect to DB configured value of times and then really shuts down. We have a redundant setup with three database hosts and switching from one db host to another db host took more than 10 seconds and mgt server got killed since it cant connect to db anymore. i have to restart the mgt server to bring it back up. To avoid this I want to configure the timeout value as well as the number of retries to get a connection -- Thanks and regards Rakesh venkatesh
Re: Shutdown VM
Disable the HA for the vm by changing filed in vm_instance table and shutdown the vm. The cloudstack wont restart it. That worked for me On Fri, Apr 5, 2019 at 1:25 PM Alex Adati wrote: > Hello, > > I'm using a computational offering with HA enabled, and would like to know > if there is any way to shut down the VM by the operating system. > > Thank's. > > Em sex, 5 de abr de 2019 às 04:54, Riepl, Gregor (SWISS TXT) < > gregor.ri...@swisstxt.ch> escreveu: > > > > You should be using a computer offering with HA enabled. > > > When you’re using this option, cloudstack will start vm until you > > shutdown it by cloudstack interface (UI or api). > > > > I think you meant: You should NOT be using a compute offering with HA > > enabled, right? > > > > > -- > -- Thanks and regards Rakesh venkatesh
Re: Network metrics
Hello Gregor You can use "libvirt exporter" to export all stats like cpu utilization, memory consumption, disk read/write, network read/write and export it to prometheus. From prometheus you can use grafana or other monitoring tools to see all stats. We use exporter merger third party tool to merge all the data from all the hypervisors and show it on dashboard. This way if any of your vm is having issues, you can quickly look at graphs and troubleshoot it. On Fri, Apr 5, 2019 at 10:01 AM Riepl, Gregor (SWISS TXT) < gregor.ri...@swisstxt.ch> wrote: > Hi, > > I recently discovered that CloudStack exposes certain VM metrics like disk > I/O, network bandwidth, CPU load and (if the HV agent is installed) memory > usage. > Presumably, these are fetched directly from the hypervisor. > > However: I found no way to get the same statistics for system VMs, in > particular virtual routers. > Does the CloudStack API offer a way to obtain per-network metrics? > > Regards, > Gregor -- Thanks and regards Rakesh venkatesh
Calling cloudstack container service API's from Cloudstack
Hello Folks I am wondering if there is a way to call cloudstack container service API from cloudstack? I want to delete all the containers created by a domain when the domain is deleted. I am using cloudstack version 4.7 with KVM and when I delete a domain, the containers created by domain are still active. So I want to call deleteContainerCluster API from cloudstack deleteDomain API. What modifications should I make in server/src/com/cloud/user/DomainManagerImpl.java of Cloudstack or DeleteContainerClusterCmd.java of CCS? Any examples of this usage are helpful. -- Thanks and regards Rakesh venkatesh
Re: Both routers are in Master state in VPC
Hi Rohit Thanks for the reply. We saw this issue in our customer setup. we are using kvm hypervisor and cloudstack 4.7.1 I will have a look at the issue you provided. On Thu, Dec 6, 2018 at 3:28 PM Rohit Yadav wrote: > Hi Rakesh, > > > Can you share your configuration/setup of VPC, hypervisor, are you using > private gateways and/or VPN with your VPCs and the version of CloudStack? > > > There are few known cases where both VPC redundant VRs become master, see: > https://github.com/apache/cloudstack/pull/2128 and CLOUDSTACK-9885 issue. > > > - Rohit > > <https://cloudstack.apache.org> > > > > > From: Rakesh Venkatesh > Sent: Thursday, December 6, 2018 4:53:31 PM > To: users@cloudstack.apache.org > Subject: Both routers are in Master state in VPC > > Hello Folks > > Several times we have encountered that both the master and the backup > router of VPC network will be in MASTER state and this causes several > issues to network connectivity. We dont the trigger for it but can be > solved by restarting the network with cleanup option. Anyone else has > encountered this issue before and know how to fix it? > > rohit.ya...@shapeblue.com > www.shapeblue.com > Amadeus House, Floral Street, London WC2E 9DPUK > @shapeblue > > > > -- Thanks and regards Rakesh venkatesh
Both routers are in Master state in VPC
Hello Folks Several times we have encountered that both the master and the backup router of VPC network will be in MASTER state and this causes several issues to network connectivity. We dont the trigger for it but can be solved by restarting the network with cleanup option. Anyone else has encountered this issue before and know how to fix it?
Re: VRs swapping with 256 MB RAM
Hello Rene Even for VR's running on KVM, 256MB is really less. Thats why we offer extra VR with 2 cores and 1GB RAM as another option so that customers can use it instead of the default 256MB and 1 core. On Tue, Sep 18, 2018 at 5:56 PM Rene Moser wrote: > Hi > > While running test for a 4.11.1 (VMware) upgrade in our lab, we run into > low memory / swapping of VRs having 256 MB RAM. After 2-3 days it became > critical because the management server connections to VRs took very > long, minutes, this resulted in many more problems all over. > > Make sure your VRs have enough RAM. > > Regards > René > -- Thanks and regards Rakesh venkatesh
Re: Rename an instance
Hello Ugo You can change the display name of a vm through database or cloudmonkey without stopping the vm. You need to change the name in two tables if you are doing through database Change display_name in "vm_instance" and "user_vm" tables. We did it for our customers and didnt face any issue On Wed, Sep 19, 2018 at 12:28 PM Ugo Vasi wrote: > Hi all, > from the cloudstack web interface (version 4.11.1 in my case) I can only > change the "Display name" by stopping the instance and clicking the icon. > > Is it possible to change the instance name? > > Is it dangerous to change the field in the database? > > Thanks in advance > -- > > *Ugo Vasi* / System Administrator > ugo.v...@procne.it <mailto:ugo.v...@procne.it> > > > > > *Procne S.r.l.* > +39 0432 486 523 > via Cotonificio, 45 > 33010 Tavagnacco (UD) > www.procne.it <http://www.procne.it/> > > > Le informazioni contenute nella presente comunicazione ed i relativi > allegati possono essere riservate e sono, comunque, destinate > esclusivamente alle persone od alla Società sopraindicati. La > diffusione, distribuzione e/o copiatura del documento trasmesso da parte > di qualsiasi soggetto diverso dal destinatario è proibita sia ai sensi > dell'art. 616 c.p., che ai sensi del Decreto Legislativo n. 196/2003 > "Codice in materia di protezione dei dati personali". Se avete ricevuto > questo messaggio per errore, vi preghiamo di distruggerlo e di informare > immediatamente Procne S.r.l. scrivendo all' indirizzo e-mail > i...@procne.it <mailto:i...@procne.it>. > > -- Thanks and regards Rakesh venkatesh
Re: Disk controller type
I think you can change the controller type in XML file of the VM but not sure if that works or not. virsh dumpxml virsh edit change the controller type, save and start the vm again On Thu, Sep 20, 2018 at 9:59 AM Adam Witwicki wrote: > Hello, > > I have a number of of instances using the wrong controller type (ATA) does > anyone know where this is set in the database so I can start them with the > VirtIO scsi controller? > > Thanks > > Adam > > > > > Disclaimer Notice: > This email has been sent by Oakford Technology Limited, while we have > checked this e-mail and any attachments for viruses, we can not guarantee > that they are virus-free. You must therefore take full responsibility for > virus checking. > This message and any attachments are confidential and should only be read > by those to whom they are addressed. If you are not the intended recipient, > please contact us, delete the message from your computer and destroy any > copies. Any distribution or copying without our prior permission is > prohibited. > Internet communications are not always secure and therefore Oakford > Technology Limited does not accept legal responsibility for this message. > The recipient is responsible for verifying its authenticity before acting > on the contents. Any views or opinions presented are solely those of the > author and do not necessarily represent those of Oakford Technology Limited. > Registered address: Oakford Technology Limited, 10 Prince Maurice Court, > Devizes, Wiltshire. SN10 2RT. > Registered in England and Wales No. 5971519 > > -- Thanks and regards Rakesh venkatesh
Re: Error while migrating vm in ACS 4.7.1
Hello Dag I have seen this issue in two scenarios 1. While migrating multiple vm's 2. When a host is put into maintenance mode So along with this vm there are other vm's that are being migrated as well. I also see this in the log which is confusing me |2018-08-10 12:17:26,030 INFO [resource.wrapper.LibvirtMigrateCommandWrapper] (agentRequest-Handler-4:null) Migration thread for i-884-14365-VM is done Does that answer your question? On Fri, Aug 10, 2018 at 4:21 PM Dag Sonstebo wrote: > Hi Rakesh, > > “There's a migration process in progress” > have you done some digging to > find out what is actually running? > > Regards, > Dag Sonstebo > Cloud Architect > ShapeBlue > > > dag.sonst...@shapeblue.com > www.shapeblue.com > 53 Chandos Place, Covent Garden, London WC2N 4HSUK > @shapeblue > > > > On 10/08/2018, 15:16, "Rakesh Venkatesh" > wrote: > > Hello Folks > > We are seeing some errors consistently while migrating vm's from one > node > another node but still not able to find out the root cause of it. Can > anyone please let me know what might be causing the issue? The vm's are > migrated successfully though > > This issue happens when I try to migrate the vm either from cloudmonkey > command or through UI. > > Management server: > > $ lsb_release -a > No LSB modules are available. > Distributor ID: Ubuntu > Description: Ubuntu 12.04.3 LTS > Release: 12.04 > Codename: precise > > > Error logs from management server: > > Aug 10 12:35:48 node13 sh[50956]: 2018-08-10 12:35:48,243{GMT} INFO [ > resource.wrapper.LibvirtMigrateCommandWrapper] > (agentRequest-Handler-5:) > Waiting for migration of i-884-15037-VM to complete, waited 3000ms > Aug 10 12:35:48 node13 sh[50956]: 2018-08-10 12:35:48,486{GMT} INFO [ > resource.wrapper.LibvirtMigrateCommandWrapper] > (agentRequest-Handler-1:) > Waiting for migration of i-884-14365-VM to complete, waited 1ms > Aug 10 12:35:49 node13 sh[50956]: libvirt: QEMU Driver error : internal > error: unable to execute QEMU command 'migrate-set-capabilities': > There's a > migration process in progressAug 10 12:35:49 node13 sh[50956]: > 2018-08-10 > 12:35:49,087{GMT} INFO [resource.wrapper.LibvirtMigrateCommandWrapper] > (agentRequest-Handler-1:) Migration thread for i-884-14365-VM is done > Aug 10 12:35:49 node13 sh[50956]: 2018-08-10 12:35:49,087{GMT} DEBUG > [resource.wrapper.LibvirtMigrateCommandWrapper] > (agentRequest-Handler-1:) > Failed to execute while migrating domain: org.libvirt.LibvirtException: > internal error: unable to execute QEMU command > 'migrate-set-capabilities': > There's a migration process in progress > Aug 10 12:35:49 node13 sh[50956]: 2018-08-10 12:35:49,087{GMT} DEBUG > [kvm. > resource.LibvirtComputingResource] (agentRequest-Handler-1:) Restoring > the > metadata of vm snapshots of vm i-884-14365-VM > Aug 10 12:35:49 node13 sh[50956]: 2018-08-10 12:35:49,088{GMT} DEBUG > [cloud. > agent.Agent] (agentRequest-Handler-1:) Seq 367-5848768541070742315: { > Ans: , > MgmtId: 48674822042491, via: 367, Ver: v1, Flags: 10, [{ > > "com.cloud.agent.api.MigrateAnswer":{"result":false,"details":"org.libvirt.LibvirtException: > internal error: unable to execute QEMU command > 'migrate-set-capabilities': > There's a migration process in progress","wait":0}}] } > > > Agent: > > # lsb_release -a No LSB modules are available. Distributor ID: Ubuntu > Description: Ubuntu 16.04.3 LTS Release: 16.04 Codename: xenial > > Error logs from agent: > > > 2018-08-10 12:17:26,030 DEBUG > [resource.wrapper.LibvirtMigrateCommandWrapper] > (agentRequest-Handler-4:null) Failed to execute while migrating domain: > org.libvirt.LibvirtException: internal error: unable to execute QEMU > command 'migrate-set-capabilities': There's a migration process in > progress > 2018-08-10 12:17:26,030 DEBUG [kvm.resource.LibvirtComputingResource] > (agentRequest-Handler-4:null) Restoring the metadata of vm snapshots > of vm > i-884-14365-VM 2018-08-10 12:17:26,031 DEBUG [cloud.agent.Agent] > (agentRequest-Handler-4:null) Seq 367-5848768541070742169: { Ans: , > MgmtId: > 48674822042491, via: 367, Ver: v1, Flags: 10, > > [{"com.cloud.agent.api.MigrateAnswer":{"result":false,"details":"org.libvirt.LibvirtException: > internal error: unable to execute QEMU command > 'migrate-set-capabilities': > There's a migration process in progress","wait":0}}] } > > 2018-08-10 12:17:26,564 INFO > [resource.wrapper.LibvirtMigrateCommandWrapper] > (agentRequest-Handler-4:null) Live migration of instance i-884-14365-VM > initiated > > > -- > Thanks and regards > Rakesh venkatesh > > > -- Thanks and regards Rakesh venkatesh
Error while migrating vm in ACS 4.7.1
Hello Folks We are seeing some errors consistently while migrating vm's from one node another node but still not able to find out the root cause of it. Can anyone please let me know what might be causing the issue? The vm's are migrated successfully though This issue happens when I try to migrate the vm either from cloudmonkey command or through UI. Management server: $ lsb_release -a No LSB modules are available. Distributor ID: Ubuntu Description: Ubuntu 12.04.3 LTS Release: 12.04 Codename: precise Error logs from management server: Aug 10 12:35:48 node13 sh[50956]: 2018-08-10 12:35:48,243{GMT} INFO [ resource.wrapper.LibvirtMigrateCommandWrapper] (agentRequest-Handler-5:) Waiting for migration of i-884-15037-VM to complete, waited 3000ms Aug 10 12:35:48 node13 sh[50956]: 2018-08-10 12:35:48,486{GMT} INFO [ resource.wrapper.LibvirtMigrateCommandWrapper] (agentRequest-Handler-1:) Waiting for migration of i-884-14365-VM to complete, waited 1ms Aug 10 12:35:49 node13 sh[50956]: libvirt: QEMU Driver error : internal error: unable to execute QEMU command 'migrate-set-capabilities': There's a migration process in progressAug 10 12:35:49 node13 sh[50956]: 2018-08-10 12:35:49,087{GMT} INFO [resource.wrapper.LibvirtMigrateCommandWrapper] (agentRequest-Handler-1:) Migration thread for i-884-14365-VM is done Aug 10 12:35:49 node13 sh[50956]: 2018-08-10 12:35:49,087{GMT} DEBUG [resource.wrapper.LibvirtMigrateCommandWrapper] (agentRequest-Handler-1:) Failed to execute while migrating domain: org.libvirt.LibvirtException: internal error: unable to execute QEMU command 'migrate-set-capabilities': There's a migration process in progress Aug 10 12:35:49 node13 sh[50956]: 2018-08-10 12:35:49,087{GMT} DEBUG [kvm. resource.LibvirtComputingResource] (agentRequest-Handler-1:) Restoring the metadata of vm snapshots of vm i-884-14365-VM Aug 10 12:35:49 node13 sh[50956]: 2018-08-10 12:35:49,088{GMT} DEBUG [cloud. agent.Agent] (agentRequest-Handler-1:) Seq 367-5848768541070742315: { Ans: , MgmtId: 48674822042491, via: 367, Ver: v1, Flags: 10, [{ "com.cloud.agent.api.MigrateAnswer":{"result":false,"details":"org.libvirt.LibvirtException: internal error: unable to execute QEMU command 'migrate-set-capabilities': There's a migration process in progress","wait":0}}] } Agent: # lsb_release -a No LSB modules are available. Distributor ID: Ubuntu Description: Ubuntu 16.04.3 LTS Release: 16.04 Codename: xenial Error logs from agent: 2018-08-10 12:17:26,030 DEBUG [resource.wrapper.LibvirtMigrateCommandWrapper] (agentRequest-Handler-4:null) Failed to execute while migrating domain: org.libvirt.LibvirtException: internal error: unable to execute QEMU command 'migrate-set-capabilities': There's a migration process in progress 2018-08-10 12:17:26,030 DEBUG [kvm.resource.LibvirtComputingResource] (agentRequest-Handler-4:null) Restoring the metadata of vm snapshots of vm i-884-14365-VM 2018-08-10 12:17:26,031 DEBUG [cloud.agent.Agent] (agentRequest-Handler-4:null) Seq 367-5848768541070742169: { Ans: , MgmtId: 48674822042491, via: 367, Ver: v1, Flags: 10, [{"com.cloud.agent.api.MigrateAnswer":{"result":false,"details":"org.libvirt.LibvirtException: internal error: unable to execute QEMU command 'migrate-set-capabilities': There's a migration process in progress","wait":0}}] } 2018-08-10 12:17:26,564 INFO [resource.wrapper.LibvirtMigrateCommandWrapper] (agentRequest-Handler-4:null) Live migration of instance i-884-14365-VM initiated -- Thanks and regards Rakesh venkatesh