Hello community, here is the log from the commit of package virt-manager for openSUSE:Factory checked in at 2019-01-10 15:23:46 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Comparing /work/SRC/openSUSE:Factory/virt-manager (Old) and /work/SRC/openSUSE:Factory/.virt-manager.new.28833 (New) ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
Package is "virt-manager" Thu Jan 10 15:23:46 2019 rev:188 rq:664194 version:2.0.0 Changes: -------- --- /work/SRC/openSUSE:Factory/virt-manager/virt-manager.changes 2018-12-04 20:55:38.948764704 +0100 +++ /work/SRC/openSUSE:Factory/.virt-manager.new.28833/virt-manager.changes 2019-01-10 15:23:48.214301756 +0100 @@ -1,0 +2,13 @@ +Wed Jan 9 15:55:22 MST 2019 - carn...@suse.com + +- Increase the timeout for a vm to start before assuming something + is wrong. (bsc#1027942) + c30b3bc6-increase-timeout-for-vm-to-start.patch + +------------------------------------------------------------------- +Mon Dec 17 14:19:11 MST 2018 - carn...@suse.com + +- fate#326698: Add support for Xen's PVH guest type + virtinst-add-pvh-support.patch + +------------------------------------------------------------------- New: ---- c30b3bc6-increase-timeout-for-vm-to-start.patch virtinst-add-pvh-support.patch ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Other differences: ------------------ ++++++ virt-manager.spec ++++++ --- /var/tmp/diff_new_pack.nTAbOt/_old 2019-01-10 15:23:48.906301041 +0100 +++ /var/tmp/diff_new_pack.nTAbOt/_new 2019-01-10 15:23:48.910301036 +0100 @@ -1,7 +1,7 @@ # # spec file for package virt-manager # -# Copyright (c) 2018 SUSE LINUX GmbH, Nuernberg, Germany. +# Copyright (c) 2019 SUSE LINUX GmbH, Nuernberg, Germany. # # All modifications and additions to the file contributed by third parties # remain the property of their copyright owners, unless otherwise agreed @@ -35,6 +35,7 @@ Patch1: b8aff280-virtinst-quickfix-ubuntu-net-preseed-insert-cdrom-error.patch Patch2: 25b88733-urldetect-Dont-overload-suse_content-variable.patch Patch3: 9308bae3-util-Fix-typo-vpcu-vcpu.patch +Patch4: c30b3bc6-increase-timeout-for-vm-to-start.patch # SUSE Only Patch70: virtman-desktop.patch Patch71: virtman-kvm.patch @@ -54,6 +55,7 @@ Patch125: virtinst-s390x-disable-graphics.patch Patch126: virtinst-add-caasp-support.patch Patch127: virtinst-add-sle15-detection-support.patch +Patch128: virtinst-add-pvh-support.patch # Bug Fixes Patch150: virtman-prevent-double-click-starting-vm-twice.patch Patch151: virtman-increase-setKeepAlive-count.patch @@ -163,6 +165,7 @@ %patch1 -p1 %patch2 -p1 %patch3 -p1 +%patch4 -p1 # SUSE Only %patch70 -p1 %patch71 -p1 @@ -182,6 +185,7 @@ %patch125 -p1 %patch126 -p1 %patch127 -p1 +%patch128 -p1 # Bug Fixes %patch150 -p1 %patch151 -p1 ++++++ c30b3bc6-increase-timeout-for-vm-to-start.patch ++++++ Subject: Update create.py From: kalebskeithley kkeit...@redhat.com Thu Nov 1 09:18:34 2018 -0400 Date: Thu Nov 15 12:05:36 2018 -0500: Git: c30b3bc6119aeb3d32e0e899a581c3cafd7aab8b 100 is a bit too short for managing/creating vms over the corporate VPN. About half the time it times out and I have to exit the virtmgr app and restart it to see the new vm. diff --git a/virtManager/create.py b/virtManager/create.py index a60e9a3f..d3844a52 100644 --- a/virtManager/create.py +++ b/virtManager/create.py @@ -2123,7 +2123,7 @@ class vmmCreate(vmmGObjectUI): self.conn.schedule_priority_tick(pollvm=True) count = 0 foundvm = None - while count < 100: + while count < 200: for vm in self.conn.list_vms(): if vm.get_uuid() == guest.uuid: foundvm = vm ++++++ virtinst-add-pvh-support.patch ++++++ References: fate#326698 - Add pvh support to virt-manager Index: virt-manager-2.0.0/virtinst/domain/os.py =================================================================== --- virt-manager-2.0.0.orig/virtinst/domain/os.py +++ virt-manager-2.0.0/virtinst/domain/os.py @@ -33,6 +33,8 @@ class DomainOs(XMLBuilder): return self.os_type == "hvm" def is_xenpv(self): return self.os_type in ["xen", "linux"] + def is_xenpvh(self): + return self.os_type in ["xenpvh", "linux"] def is_container(self): return self.os_type == "exe" Index: virt-manager-2.0.0/virtinst/guest.py =================================================================== --- virt-manager-2.0.0.orig/virtinst/guest.py +++ virt-manager-2.0.0/virtinst/guest.py @@ -565,7 +565,7 @@ class Guest(XMLBuilder): usb_tablet = False usb_keyboard = False - if self.os.is_x86() and not self.os.is_xenpv(): + if self.os.is_x86() and not self.os.is_xenpv() and not self.os.is_xenpvh(): usb_tablet = self.osinfo.supports_usbtablet() if self.os.is_arm_machvirt(): usb_tablet = True Index: virt-manager-2.0.0/virtManager/domain.py =================================================================== --- virt-manager-2.0.0.orig/virtManager/domain.py +++ virt-manager-2.0.0/virtManager/domain.py @@ -1207,6 +1207,8 @@ class vmmDomain(vmmLibvirtObject): return self.get_xmlobj().os.is_xenpv() def is_hvm(self): return self.get_xmlobj().os.is_hvm() + def is_xenpvh(self): + return self.get_xmlobj().os.is_xenpvh() def get_uuid(self): if self._uuid is None: Index: virt-manager-2.0.0/virtManager/connection.py =================================================================== --- virt-manager-2.0.0.orig/virtManager/connection.py +++ virt-manager-2.0.0/virtManager/connection.py @@ -229,6 +229,8 @@ class vmmConnection(vmmGObject): label = "xen (paravirt)" elif gtype == "hvm": label = "xen (fullvirt)" + elif gtype == "xenpvh": + label = "xen (pvh - Technical Preview)" elif domtype == "test": if gtype == "xen": label = "test (xen)" ++++++ virtman-increase-setKeepAlive-count.patch ++++++ --- /var/tmp/diff_new_pack.nTAbOt/_old 2019-01-10 15:23:49.078300863 +0100 +++ /var/tmp/diff_new_pack.nTAbOt/_new 2019-01-10 15:23:49.082300859 +0100 @@ -6,7 +6,7 @@ =================================================================== --- virt-manager-2.0.0.orig/virtManager/connection.py +++ virt-manager-2.0.0/virtManager/connection.py -@@ -1059,7 +1059,7 @@ class vmmConnection(vmmGObject): +@@ -1061,7 +1061,7 @@ class vmmConnection(vmmGObject): self.caps.get_cpu_values(self.caps.host.cpu.arch) try: