Many thanks to Yuval.

After moving the discussion to #ovirt, I tried "fstrim -a" and this allowed the upgrade to complete successfully.

Matt







On 07/03/2018 12:19 PM, Yuval Turgeman wrote:
Hi Matt,

I would try to run `fstrim -a` (man fstrim) and see if it frees anything from the thinpool.  If you do decide to run this, please send the output for lvs again.

Also, are you on #ovirt ?

Thanks,
Yuval.


On Tue, Jul 3, 2018 at 9:00 PM, Matt Simonsen <m...@khoza.com <mailto:m...@khoza.com>> wrote:

    Thank you again for the assistance with this issue.

    Below is the result of the command below.

    In the future I am considering using different Logical RAID
    Volumes to get different devices (sda, sdb, etc) for the oVirt
    Node image & storage filesystem to simplify.  However I'd like to
    understand why this upgrade failed and also how to correct it if
    at all possible.

    I believe I need to recreate the /var/crash partition? I
    incorrectly removed it, is it simply a matter of using LVM to add
    a new partition and format it?

    Secondly, do you have any suggestions on how to move forward with
    the error regarding the pool capacity? I'm not sure if this is a
    legitimate error or problem in the upgrade process.

    Thanks,

    Matt




    On 07/03/2018 03:58 AM, Yuval Turgeman wrote:
    Not sure this is the problem, autoextend should be enabled for
    the thinpool, `lvs -o +profile` should show imgbased-pool
    (defined at /etc/lvm/profile/imgbased-pool.profile)

    On Tue, Jul 3, 2018 at 8:55 AM, Yedidyah Bar David
    <d...@redhat.com <mailto:d...@redhat.com>> wrote:

        On Mon, Jul 2, 2018 at 7:54 PM, Matt Simonsen <m...@khoza.com
        <mailto:m...@khoza.com>> wrote:
        >
        > This error adds some clarity.
        >
        > That said, I'm a bit unsure how the space can be the issue
        given I have several hundred GB of storage in the thin pool
        that's unused...
        >
        > How do you suggest I proceed?
        >
        > Thank you for your help,
        >
        > Matt
        >
        >
        >
        > [root@node6-g8-h4 ~]# lvs
        >
        >   LV  VG              Attr       LSize   Pool  Origin     
                               Data% Meta%  Move Log Cpy%Sync Convert
        >   home  onn_node1-g8-h4 Vwi-aotz--   1.00g pool00          
                                4.79
        >   ovirt-node-ng-4.2.2-0.20180423.0    onn_node1-g8-h4
        Vwi---tz-k <50.06g pool00 root
        >   ovirt-node-ng-4.2.2-0.20180423.0+1  onn_node1-g8-h4
        Vwi---tz-- <50.06g pool00 ovirt-node-ng-4.2.2-0.20180423.0
        >   ovirt-node-ng-4.2.3.1-0.20180530.0  onn_node1-g8-h4
        Vri---tz-k <50.06g pool00
        >   ovirt-node-ng-4.2.3.1-0.20180530.0+1 onn_node1-g8-h4
        Vwi-aotz-- <50.06g pool00 ovirt-node-ng-4.2.3.1-0.20180530.0 6.95
        >   pool00  onn_node1-g8-h4 twi-aotz--  <1.30t              
                               76.63 50.34

        I think your thinpool meta volume is close to full and needs
        to be enlarged.
        This quite likely happened because you extended the thinpool
        without
        extending the meta vol.

        Check also 'lvs -a'.

        This might be enough, but check the names first:

        lvextend -L+200m onn_node1-g8-h4/pool00_tmeta

        Best regards,

        >   root  onn_node1-g8-h4 Vwi---tz-- <50.06g pool00
        >   tmp   onn_node1-g8-h4 Vwi-aotz--   1.00g pool00 5.04
        >   var   onn_node1-g8-h4 Vwi-aotz--  15.00g pool00 5.86
        >   var_crash   onn_node1-g8-h4 Vwi---tz--  10.00g pool00
        >   var_local_images  onn_node1-g8-h4 Vwi-aotz--   1.10t
        pool00 89.72
        >   var_log   onn_node1-g8-h4 Vwi-aotz--   8.00g pool00 6.84
        >   var_log_audit   onn_node1-g8-h4 Vwi-aotz--   2.00g pool00
        6.16
        > [root@node6-g8-h4 ~]# vgs
        >   VG              #PV #LV #SN Attr  VSize  VFree
        >   onn_node1-g8-h4   1  13   0 wz--n- <1.31t 8.00g
        >
        >
        > 2018-06-29 14:19:31,142 [DEBUG] (MainThread) Version:
        imgbased-1.0.20
        > 2018-06-29 14:19:31,147 [DEBUG] (MainThread) Arguments:
        
Namespace(FILENAME='/usr/share/ovirt-node-ng/image//ovirt-node-ng-4.2.0-0.20180626.0.el7.squashfs.img',
        command='update', debug=True, experimental=False,
        format='liveimg', stream='Image')
        > 2018-06-29 14:19:31,147 [INFO] (MainThread) Extracting
        image
        
'/usr/share/ovirt-node-ng/image//ovirt-node-ng-4.2.0-0.20180626.0.el7.squashfs.img'
        > 2018-06-29 14:19:31,148 [DEBUG] (MainThread) Calling
        binary: (['mktemp', '-d', '--tmpdir', 'mnt.XXXXX'],) {}
        > 2018-06-29 14:19:31,148 [DEBUG] (MainThread) Calling:
        (['mktemp', '-d', '--tmpdir', 'mnt.XXXXX'],) {'close_fds':
        True, 'stderr': -2}
        > 2018-06-29 14:19:31,150 [DEBUG] (MainThread) Returned:
        /tmp/mnt.1OhaU
        > 2018-06-29 14:19:31,151 [DEBUG] (MainThread) Calling
        binary: (['mount',
        
'/usr/share/ovirt-node-ng/image//ovirt-node-ng-4.2.0-0.20180626.0.el7.squashfs.img',
        u'/tmp/mnt.1OhaU'],) {}
        > 2018-06-29 14:19:31,151 [DEBUG] (MainThread) Calling:
        (['mount',
        
'/usr/share/ovirt-node-ng/image//ovirt-node-ng-4.2.0-0.20180626.0.el7.squashfs.img',
        u'/tmp/mnt.1OhaU'],) {'close_fds': True, 'stderr': -2}
        > 2018-06-29 14:19:31,157 [DEBUG] (MainThread) Returned:
        > 2018-06-29 14:19:31,158 [DEBUG] (MainThread) Mounted squashfs
        > 2018-06-29 14:19:31,158 [DEBUG] (MainThread) Found fsimage
        at '/tmp/mnt.1OhaU/LiveOS/rootfs.img'
        > 2018-06-29 14:19:31,159 [DEBUG] (MainThread) Calling
        binary: (['mktemp', '-d', '--tmpdir', 'mnt.XXXXX'],) {}
        > 2018-06-29 14:19:31,159 [DEBUG] (MainThread) Calling:
        (['mktemp', '-d', '--tmpdir', 'mnt.XXXXX'],) {'close_fds':
        True, 'stderr': -2}
        > 2018-06-29 14:19:31,162 [DEBUG] (MainThread) Returned:
        /tmp/mnt.153do
        > 2018-06-29 14:19:31,162 [DEBUG] (MainThread) Calling
        binary: (['mount', u'/tmp/mnt.1OhaU/LiveOS/rootfs.img',
        u'/tmp/mnt.153do'],) {}
        > 2018-06-29 14:19:31,162 [DEBUG] (MainThread) Calling:
        (['mount', u'/tmp/mnt.1OhaU/LiveOS/rootfs.img',
        u'/tmp/mnt.153do'],) {'close_fds': True, 'stderr': -2}
        > 2018-06-29 14:19:31,177 [DEBUG] (MainThread) Returned:
        > 2018-06-29 14:19:31,189 [DEBUG] (MainThread) Using nvr:
        ovirt-node-ng-4.2.4-0.20180626.0
        > 2018-06-29 14:19:31,189 [DEBUG] (MainThread) Fetching image
        for '/'
        > 2018-06-29 14:19:31,189 [DEBUG] (MainThread) Calling
        binary: (['findmnt', '--noheadings', '-o', 'SOURCE', '/'],) {}
        > 2018-06-29 14:19:31,190 [DEBUG] (MainThread) Calling:
        (['findmnt', '--noheadings', '-o', 'SOURCE', '/'],)
        {'close_fds': True, 'stderr': -2}
        > 2018-06-29 14:19:31,203 [DEBUG] (MainThread) Returned:
        /dev/mapper/onn_node1--g8--h4-ovirt--node--ng--4.2.3.1--0.20180530.0+1
        > 2018-06-29 14:19:31,204 [DEBUG] (MainThread) Found
        '/dev/mapper/onn_node1--g8--h4-ovirt--node--ng--4.2.3.1--0.20180530.0+1'
        > 2018-06-29 14:19:31,204 [DEBUG] (MainThread) Calling
        binary: (['lvs', '--noheadings', '--ignoreskippedcluster',
        '-ovg_name,lv_name',
        
u'/dev/mapper/onn_node1--g8--h4-ovirt--node--ng--4.2.3.1--0.20180530.0+1'],)
        {'stderr': <open file '/dev/null', mode 'w' at 0x7f56b787eed0>}
        > 2018-06-29 14:19:31,204 [DEBUG] (MainThread) Calling:
        (['lvs', '--noheadings', '--ignoreskippedcluster',
        '-ovg_name,lv_name',
        
u'/dev/mapper/onn_node1--g8--h4-ovirt--node--ng--4.2.3.1--0.20180530.0+1'],)
        {'close_fds': True, 'stderr': <open file '/dev/null', mode
        'w' at 0x7f56b787eed0>}
        > 2018-06-29 14:19:31,283 [DEBUG] (MainThread) Returned:
        onn_node1-g8-h4 ovirt-node-ng-4.2.3.1-0.20180530.0+1
        > 2018-06-29 14:19:31,283 [DEBUG] (MainThread) Found LV for
        path
        /dev/mapper/onn_node1--g8--h4-ovirt--node--ng--4.2.3.1--0.20180530.0+1:
        onn_node1-g8-h4 ovirt-node-ng-4.2.3.1-0.20180530.0+1
        > 2018-06-29 14:19:31,283 [DEBUG] (MainThread) Found LV
        'ovirt-node-ng-4.2.3.1-0.20180530.0+1' for path
        '/dev/mapper/onn_node1--g8--h4-ovirt--node--ng--4.2.3.1--0.20180530.0+1'
        > 2018-06-29 14:19:31,284 [DEBUG] (MainThread) Calling
        binary: (['vgs', '--noheadings', '--ignoreskippedcluster',
        '--select', 'vg_tags = imgbased:vg', '-o', 'vg_name'],)
        {'stderr': <open file '/dev/null', mode 'w' at 0x7f56b787eed0>}
        > 2018-06-29 14:19:31,284 [DEBUG] (MainThread) Calling:
        (['vgs', '--noheadings', '--ignoreskippedcluster',
        '--select', 'vg_tags = imgbased:vg', '-o', 'vg_name'],)
        {'close_fds': True, 'stderr': <open file '/dev/null', mode
        'w' at 0x7f56b787eed0>}
        > 2018-06-29 14:19:31,321 [DEBUG] (MainThread) Returned:
        onn_node1-g8-h4
        > 2018-06-29 14:19:31,322 [DEBUG] (MainThread) Calling
        binary: (['lvs', '--noheadings', '--ignoreskippedcluster',
        '-osize', '--units', 'B',
        u'onn_node1-g8-h4/ovirt-node-ng-4.2.3.1-0.20180530.0+1'],)
        {'stderr': <open file '/dev/null', mode 'w' at 0x7f56b787eed0>}
        > 2018-06-29 14:19:31,322 [DEBUG] (MainThread) Calling:
        (['lvs', '--noheadings', '--ignoreskippedcluster', '-osize',
        '--units', 'B',
        u'onn_node1-g8-h4/ovirt-node-ng-4.2.3.1-0.20180530.0+1'],)
        {'close_fds': True, 'stderr': <open file '/dev/null', mode
        'w' at 0x7f56b787eed0>}
        > 2018-06-29 14:19:31,355 [DEBUG] (MainThread) Returned:
        53750005760B
        > 2018-06-29 14:19:31,355 [DEBUG] (MainThread) Recommeneded
        base size: 53750005760B
        > 2018-06-29 14:19:31,355 [INFO] (MainThread) Starting base
        creation
        > 2018-06-29 14:19:31,355 [INFO] (MainThread) New base will
        be: ovirt-node-ng-4.2.4-0.20180626.0
        > 2018-06-29 14:19:31,356 [DEBUG] (MainThread) Calling
        binary: (['vgs', '--noheadings', '--ignoreskippedcluster',
        '@imgbased:pool', '-o', 'lv_full_name'],) {'stderr': <open
        file '/dev/null', mode 'w' at 0x7f56b787eed0>}
        > 2018-06-29 14:19:31,356 [DEBUG] (MainThread) Calling:
        (['vgs', '--noheadings', '--ignoreskippedcluster',
        '@imgbased:pool', '-o', 'lv_full_name'],) {'close_fds': True,
        'stderr': <open file '/dev/null', mode 'w' at 0x7f56b787eed0>}
        > 2018-06-29 14:19:31,381 [DEBUG] (MainThread) Returned:
        onn_node1-g8-h4/pool00
        > 2018-06-29 14:19:31,381 [DEBUG] (MainThread) Pool: <LV
        'onn_node1-g8-h4/pool00' />
        > 2018-06-29 14:19:31,382 [DEBUG] (MainThread) Calling
        binary: (['lvcreate', '--thin', '--virtualsize',
        u'53750005760B', '--name',
        'ovirt-node-ng-4.2.4-0.20180626.0',
        u'onn_node1-g8-h4/pool00'],) {}
        > 2018-06-29 14:19:31,382 [DEBUG] (MainThread) Calling:
        (['lvcreate', '--thin', '--virtualsize', u'53750005760B',
        '--name', 'ovirt-node-ng-4.2.4-0.20180626.0',
        u'onn_node1-g8-h4/pool00'],) {'close_fds': True, 'stderr': -2}
        > 2018-06-29 14:19:31,406 [DEBUG] (MainThread) Exception! 
         Cannot create new thin volume, free space in thin pool
        onn_node1-g8-h4/pool00 reached threshold.
        >
        > 2018-06-29 14:19:31,406 [DEBUG] (MainThread) Calling
        binary: (['umount', '-l', u'/tmp/mnt.153do'],) {}
        > 2018-06-29 14:19:31,406 [DEBUG] (MainThread) Calling:
        (['umount', '-l', u'/tmp/mnt.153do'],) {'close_fds': True,
        'stderr': -2}
        > 2018-06-29 14:19:31,422 [DEBUG] (MainThread) Returned:
        > 2018-06-29 14:19:31,422 [DEBUG] (MainThread) Calling
        binary: (['rmdir', u'/tmp/mnt.153do'],) {}
        > 2018-06-29 14:19:31,422 [DEBUG] (MainThread) Calling:
        (['rmdir', u'/tmp/mnt.153do'],) {'close_fds': True, 'stderr': -2}
        > 2018-06-29 14:19:31,425 [DEBUG] (MainThread) Returned:
        > 2018-06-29 14:19:31,425 [DEBUG] (MainThread) Calling
        binary: (['umount', '-l', u'/tmp/mnt.1OhaU'],) {}
        > 2018-06-29 14:19:31,425 [DEBUG] (MainThread) Calling:
        (['umount', '-l', u'/tmp/mnt.1OhaU'],) {'close_fds': True,
        'stderr': -2}
        > 2018-06-29 14:19:31,437 [DEBUG] (MainThread) Returned:
        > 2018-06-29 14:19:31,437 [DEBUG] (MainThread) Calling
        binary: (['rmdir', u'/tmp/mnt.1OhaU'],) {}
        > 2018-06-29 14:19:31,437 [DEBUG] (MainThread) Calling:
        (['rmdir', u'/tmp/mnt.1OhaU'],) {'close_fds': True, 'stderr': -2}
        > 2018-06-29 14:19:31,440 [DEBUG] (MainThread) Returned:
        > Traceback (most recent call last):
        >   File "/usr/lib64/python2.7/runpy.py", line 162, in
        _run_module_as_main
        >     "__main__", fname, loader, pkg_name)
        >   File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code
        >     exec code in run_globals
        >   File
        
"/tmp/tmp.mzQBYouvWT/usr/lib/python2.7/site-packages/imgbased/__main__.py",
        line 53, in <module>
        >     CliApplication()
        >   File
        
"/tmp/tmp.mzQBYouvWT/usr/lib/python2.7/site-packages/imgbased/__init__.py",
        line 82, in CliApplication
        >     app.hooks.emit("post-arg-parse", args)
        >   File
        "/tmp/tmp.mzQBYouvWT/usr/lib/python2.7/site-packages/imgbased/hooks.py",
        line 120, in emit
        >     cb(self.context, *args)
        >   File
        
"/tmp/tmp.mzQBYouvWT/usr/lib/python2.7/site-packages/imgbased/plugins/update.py",
        line 56, in post_argparse
        >     base_lv, _ =
        LiveimgExtractor(app.imgbase).extract(args.FILENAME)
        >   File
        
"/tmp/tmp.mzQBYouvWT/usr/lib/python2.7/site-packages/imgbased/plugins/update.py",
        line 118, in extract
        >     "%s" % size, nvr)
        >   File
        
"/tmp/tmp.mzQBYouvWT/usr/lib/python2.7/site-packages/imgbased/plugins/update.py",
        line 84, in add_base_with_tree
        >     lvs)
        >   File
        
"/tmp/tmp.mzQBYouvWT/usr/lib/python2.7/site-packages/imgbased/imgbase.py",
        line 310, in add_base
        >     new_base_lv = pool.create_thinvol(new_base.lv_name, size)
        >   File
        "/tmp/tmp.mzQBYouvWT/usr/lib/python2.7/site-packages/imgbased/lvm.py",
        line 324, in create_thinvol
        >     self.lvm_name])
        >   File
        "/tmp/tmp.mzQBYouvWT/usr/lib/python2.7/site-packages/imgbased/utils.py",
        line 390, in lvcreate
        >     return self.call(["lvcreate"] + args, **kwargs)
        >   File
        "/tmp/tmp.mzQBYouvWT/usr/lib/python2.7/site-packages/imgbased/utils.py",
        line 378, in call
        >     stdout = call(*args, **kwargs)
        >   File
        "/tmp/tmp.mzQBYouvWT/usr/lib/python2.7/site-packages/imgbased/utils.py",
        line 153, in call
        >     return subprocess.check_output(*args, **kwargs).strip()
        >   File "/usr/lib64/python2.7/subprocess.py", line 575, in
        check_output
        >     raise CalledProcessError(retcode, cmd, output=output)
        > subprocess.CalledProcessError: Command '['lvcreate',
        '--thin', '--virtualsize', u'53750005760B', '--name',
        'ovirt-node-ng-4.2.4-0.20180626.0',
        u'onn_node1-g8-h4/pool00']' returned non-zero exit status 5
        >
        >
        >
        >
        >
        > On 07/02/2018 04:58 AM, Yuval Turgeman wrote:
        >
        > Looks like the upgrade script failed - can you please
        attach /var/log/imgbased.log or /tmp/imgbased.log ?
        >
        > Thanks,
        > Yuval.
        >
        > On Mon, Jul 2, 2018 at 2:54 PM, Sandro Bonazzola
        <sbona...@redhat.com <mailto:sbona...@redhat.com>> wrote:
        >>
        >> Yuval, can you please have a look?
        >>
        >> 2018-06-30 7:48 GMT+02:00 Oliver Riesener
        <oliver.riese...@hs-bremen.de
        <mailto:oliver.riese...@hs-bremen.de>>:
        >>>
        >>> Yes, here is the same.
        >>>
        >>> It seams the bootloader isn’t configured right ?
        >>>
        >>> I did the Upgrade and reboot to 4.2.4 from UI and got:
        >>>
        >>> [root@ovn-monster ~]# nodectl info
        >>> layers:
        >>>  ovirt-node-ng-4.2.4-0.20180626.0:
        >>>  ovirt-node-ng-4.2.4-0.20180626.0+1
        >>>  ovirt-node-ng-4.2.3.1-0.20180530.0:
        >>>  ovirt-node-ng-4.2.3.1-0.20180530.0+1
        >>>  ovirt-node-ng-4.2.3-0.20180524.0:
        >>>  ovirt-node-ng-4.2.3-0.20180524.0+1
        >>>  ovirt-node-ng-4.2.1.1-0.20180223.0:
        >>>  ovirt-node-ng-4.2.1.1-0.20180223.0+1
        >>> bootloader:
        >>>   default: ovirt-node-ng-4.2.3-0.20180524.0+1
        >>>   entries:
        >>>  ovirt-node-ng-4.2.3-0.20180524.0+1:
        >>>       index: 0
        >>>       title: ovirt-node-ng-4.2.3-0.20180524.0
        >>>       kernel:
        
/boot/ovirt-node-ng-4.2.3-0.20180524.0+1/vmlinuz-3.10.0-862.3.2.el7.x86_64
        >>>       args: "ro crashkernel=auto rd.lvm.lv
        <http://rd.lvm.lv>=onn_ovn-monster/ovirt-node-ng-4.2.3-0.20180524.0+1
        rd.lvm.lv <http://rd.lvm.lv>=onn_ovn-monster/swap
        rd.md.uuid=c6c3013b:027a9346:67dfd181:89635587 rhgb quiet
        LANG=de_DE.UTF-8 img.bootid=ovirt-node-ng-4.2.3-0.20180524.0+1"
        >>>       initrd:
        
/boot/ovirt-node-ng-4.2.3-0.20180524.0+1/initramfs-3.10.0-862.3.2.el7.x86_64.img
        >>>       root:
        /dev/onn_ovn-monster/ovirt-node-ng-4.2.3-0.20180524.0+1
        >>>  ovirt-node-ng-4.2.1.1-0.20180223.0+1:
        >>>       index: 1
        >>>       title: ovirt-node-ng-4.2.1.1-0.20180223.0
        >>>       kernel:
        
/boot/ovirt-node-ng-4.2.1.1-0.20180223.0+1/vmlinuz-3.10.0-693.17.1.el7.x86_64
        >>>       args: "ro crashkernel=auto rd.lvm.lv
        <http://rd.lvm.lv>=onn_ovn-monster/ovirt-node-ng-4.2.1.1-0.20180223.0+1
        rd.lvm.lv <http://rd.lvm.lv>=onn_ovn-monster/swap
        rd.md.uuid=c6c3013b:027a9346:67dfd181:89635587 rhgb quiet
        LANG=de_DE.UTF-8 img.bootid=ovirt-node-ng-4.2.1.1-0.20180223.0+1"
        >>>       initrd:
        
/boot/ovirt-node-ng-4.2.1.1-0.20180223.0+1/initramfs-3.10.0-693.17.1.el7.x86_64.img
        >>>       root:
        /dev/onn_ovn-monster/ovirt-node-ng-4.2.1.1-0.20180223.0+1
        >>> current_layer: ovirt-node-ng-4.2.3-0.20180524.0+1
        >>> [root@ovn-monster ~]# uptime
        >>>  07:35:27 up 2 days, 15:42, 1 user,  load average: 1,07,
        1,00, 0,95
        >>>
        >>> Am 29.06.2018 um 23:53 schrieb Matt Simonsen
        <m...@khoza.com <mailto:m...@khoza.com>>:
        >>>
        >>> Hello,
        >>>
        >>> I did yum updates on 2 of my oVirt 4.2.3 nodes running
        the prebuilt node platform and it doesn't appear the updates
        worked.
        >>>
        >>>
        >>> [root@node6-g8-h4 ~]# yum update
        >>> Loaded plugins: enabled_repos_upload, fastestmirror,
        imgbased-persist,
        >>>               : package_upload, product-id,
        search-disabled-repos, subscription-
        >>>               : manager
        >>> This system is not registered with an entitlement server.
        You can use subscription-manager to register.
        >>> Loading mirror speeds from cached hostfile
        >>>  * ovirt-4.2-epel: linux.mirrors.es.net
        <http://linux.mirrors.es.net>
        >>> Resolving Dependencies
        >>> --> Running transaction check
        >>> ---> Package ovirt-node-ng-image-update.noarch
        0:4.2.3.1-1.el7 will be updated
        >>> ---> Package ovirt-node-ng-image-update.noarch
        0:4.2.4-1.el7 will be obsoleting
        >>> ---> Package
        ovirt-node-ng-image-update-placeholder.noarch 0:4.2.3.1-1.el7
        will be obsoleted
        >>> --> Finished Dependency Resolution
        >>>
        >>> Dependencies Resolved
        >>>
        >>>
        
=========================================================================================================================
        >>>  Package             Arch Version    Repository         
             Size
        >>>
        
=========================================================================================================================
        >>> Installing:
        >>>  ovirt-node-ng-image-update              noarch
        4.2.4-1.el7      ovirt-4.2               647 M
        >>>      replacing
        ovirt-node-ng-image-update-placeholder.noarch 4.2.3.1-1.el7
        >>>
        >>> Transaction Summary
        >>>
        
=========================================================================================================================
        >>> Install  1 Package
        >>>
        >>> Total download size: 647 M
        >>> Is this ok [y/d/N]: y
        >>> Downloading packages:
        >>> warning:
        
/var/cache/yum/x86_64/7/ovirt-4.2/packages/ovirt-node-ng-image-update-4.2.4-1.el7.noarch.rpm:
        Header V4 RSA/SHA1 Signature, key ID fe590cb7: NOKEY
        >>> Public key for
        ovirt-node-ng-image-update-4.2.4-1.el7.noarch.rpm is not
        installed
        >>> ovirt-node-ng-image-update-4.2.4-1.el7.noarch.rpm | 647
        MB  00:02:07
        >>> Retrieving key from file:///etc/pki/rpm-gpg/RPM-GPG-ovirt-4.2
        >>> Importing GPG key 0xFE590CB7:
        >>>  Userid     : "oVirt <in...@ovirt.org
        <mailto:in...@ovirt.org>>"
        >>>  Fingerprint: 31a5 d783 7fad 7cb2 86cd 3469 ab8c 4f9d
        fe59 0cb7
        >>>  Package    : ovirt-release42-4.2.3.1-1.el7.noarch
        (installed)
        >>>  From       : /etc/pki/rpm-gpg/RPM-GPG-ovirt-4.2
        >>> Is this ok [y/N]: y
        >>> Running transaction check
        >>> Running transaction test
        >>> Transaction test succeeded
        >>> Running transaction
        >>>   Installing :
        ovirt-node-ng-image-update-4.2.4-1.el7.noarch 1/3
        >>> warning:
        %post(ovirt-node-ng-image-update-4.2.4-1.el7.noarch)
        scriptlet failed, exit status 1
        >>> Non-fatal POSTIN scriptlet failure in rpm package
        ovirt-node-ng-image-update-4.2.4-1.el7.noarch
        >>>   Erasing    :
        ovirt-node-ng-image-update-placeholder-4.2.3.1-1.el7.noarch 2/3
        >>>   Cleanup    :
        ovirt-node-ng-image-update-4.2.3.1-1.el7.noarch 3/3
        >>> warning: file
        
/usr/share/ovirt-node-ng/image/ovirt-node-ng-4.2.0-0.20180530.0.el7.squashfs.img:
        remove failed: No such file or directory
        >>> Uploading Package Profile
        >>> Unable to upload Package Profile
        >>>   Verifying  :
        ovirt-node-ng-image-update-4.2.4-1.el7.noarch 1/3
        >>>   Verifying  :
        ovirt-node-ng-image-update-4.2.3.1-1.el7.noarch 2/3
        >>>   Verifying  :
        ovirt-node-ng-image-update-placeholder-4.2.3.1-1.el7.noarch 3/3
        >>>
        >>> Installed:
        >>> ovirt-node-ng-image-update.no
        <http://ovirt-node-ng-image-update.no>arch 0:4.2.4-1.el7
        >>>
        >>> Replaced:
        >>>  ovirt-node-ng-image-update-placeholder.noarch
        0:4.2.3.1-1.el7
        >>>
        >>> Complete!
        >>> Uploading Enabled Repositories Report
        >>> Loaded plugins: fastestmirror, product-id,
        subscription-manager
        >>> This system is not registered with an entitlement server.
        You can use subscription-manager to register.
        >>> Cannot upload enabled repos report, is this client
        registered?
        >>>
        >>>
        >>> My engine shows the nodes as having no updates, however
        the major components including the kernel version and port
        9090 admin GUI show 4.2.3
        >>>
        >>> Is there anything I can provide to help diagnose the issue?
        >>>
        >>>
        >>> [root@node6-g8-h4 ~]# rpm -qa | grep ovirt
        >>>
        >>> ovirt-imageio-common-1.3.1.2-0.el7.centos.noarch
        >>> ovirt-host-deploy-1.7.3-1.el7.centos.noarch
        >>> ovirt-vmconsole-host-1.0.5-4.el7.centos.noarch
        >>> ovirt-provider-ovn-driver-1.2.10-1.el7.centos.noarch
        >>> ovirt-engine-sdk-python-3.6.9.1-1.el7.noarch
        >>> ovirt-setup-lib-1.1.4-1.el7.centos.noarch
        >>> ovirt-release42-4.2.3.1-1.el7.noarch
        >>> ovirt-imageio-daemon-1.3.1.2-0.el7.centos.noarch
        >>> ovirt-hosted-engine-setup-2.2.20-1.el7.centos.noarch
        >>> ovirt-host-dependencies-4.2.2-2.el7.centos.x86_64
        >>> ovirt-hosted-engine-ha-2.2.11-1.el7.centos.noarch
        >>> ovirt-host-4.2.2-2.el7.centos.x86_64
        >>> ovirt-node-ng-image-update-4.2.4-1.el7.noarch
        >>> ovirt-vmconsole-1.0.5-4.el7.centos.noarch
        >>> ovirt-release-host-node-4.2.3.1-1.el7.noarch
        >>> cockpit-ovirt-dashboard-0.11.24-1.el7.centos.noarch
        >>> ovirt-node-ng-nodectl-4.2.0-0.20180524.0.el7.noarch
        >>> python-ovirt-engine-sdk4-4.2.6-2.el7.centos.x86_64
        >>>
        >>> [root@node6-g8-h4 ~]# yum update
        >>> Loaded plugins: enabled_repos_upload, fastestmirror,
        imgbased-persist, package_upload, product-id,
        search-disabled-repos, subscription-manager
        >>> This system is not registered with an entitlement server.
        You can use subscription-manager to register.
        >>> Loading mirror speeds from cached hostfile
        >>>  * ovirt-4.2-epel: linux.mirrors.es.net
        <http://linux.mirrors.es.net>
        >>> No packages marked for update
        >>> Uploading Enabled Repositories Report
        >>> Loaded plugins: fastestmirror, product-id,
        subscription-manager
        >>> This system is not registered with an entitlement server.
        You can use subscription-manager to register.
        >>> Cannot upload enabled repos report, is this client
        registered?
        >>> _______________________________________________
        >>> Users mailing list -- users@ovirt.org
        <mailto:users@ovirt.org>
        >>> To unsubscribe send an email to users-le...@ovirt.org
        <mailto:users-le...@ovirt.org>
        >>> Privacy Statement:
        https://www.ovirt.org/site/privacy-policy/
        <https://www.ovirt.org/site/privacy-policy/>
        >>> oVirt Code of Conduct:
        https://www.ovirt.org/community/about/community-guidelines/
        <https://www.ovirt.org/community/about/community-guidelines/>
        >>> List Archives:
        
https://lists.ovirt.org/archives/list/users@ovirt.org/message/UHQMGULUHL4GBBHUBNGOAICJEM6W3RVW/
        
<https://lists.ovirt.org/archives/list/users@ovirt.org/message/UHQMGULUHL4GBBHUBNGOAICJEM6W3RVW/>
        >>>
        >>>
        >>>
        >>> _______________________________________________
        >>> Users mailing list -- users@ovirt.org
        <mailto:users@ovirt.org>
        >>> To unsubscribe send an email to users-le...@ovirt.org
        <mailto:users-le...@ovirt.org>
        >>> Privacy Statement:
        https://www.ovirt.org/site/privacy-policy/
        <https://www.ovirt.org/site/privacy-policy/>
        >>> oVirt Code of Conduct:
        https://www.ovirt.org/community/about/community-guidelines/
        <https://www.ovirt.org/community/about/community-guidelines/>
        >>> List Archives:
        
https://lists.ovirt.org/archives/list/users@ovirt.org/message/RTOXFNAXQ3NJBWX7RXOYK5H5RZBHX2OK/
        
<https://lists.ovirt.org/archives/list/users@ovirt.org/message/RTOXFNAXQ3NJBWX7RXOYK5H5RZBHX2OK/>
        >>>
        >>
        >>
        >>
        >> --
        >>
        >> SANDRO BONAZZOLA
        >>
        >> MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
        >>
        >> Red Hat EMEA
        >>
        >> sbona...@redhat.com <mailto:sbona...@redhat.com>
        >
        >
        >
        >
        > _______________________________________________
        > Users mailing list -- users@ovirt.org <mailto:users@ovirt.org>
        > To unsubscribe send an email to users-le...@ovirt.org
        <mailto:users-le...@ovirt.org>
        > Privacy Statement:
        https://www.ovirt.org/site/privacy-policy/
        <https://www.ovirt.org/site/privacy-policy/>
        > oVirt Code of Conduct:
        https://www.ovirt.org/community/about/community-guidelines/
        <https://www.ovirt.org/community/about/community-guidelines/>
        > List Archives:
        
https://lists.ovirt.org/archives/list/users@ovirt.org/message/AS3UWWIO5PBFYUPW5DIJ6O6VYJAL5ZIL/
        
<https://lists.ovirt.org/archives/list/users@ovirt.org/message/AS3UWWIO5PBFYUPW5DIJ6O6VYJAL5ZIL/>
        >



-- Didi





_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/GHCM7W5C25TRDZF7CMMJPM3KW55E5VZT/

Reply via email to