Hello,

I cannot flatten an image, it always restarts with:

[email protected]:~# rbd flatten vm-hdd/vm-104-disk-1
Image flatten: 28% complete...2021-04-29 10:50:27.373 7ff7caffd700 -1 librbd::operation::FlattenRequest: 0x7ff7c4009db0 should_complete: encountered error: (85) Interrupted system call should be restarted Image flatten: 26% complete...2021-04-29 10:50:33.053 7ff7caffd700 -1 librbd::operation::FlattenRequest: 0x7ff7c4008fc0 should_complete: encountered error: (85) Interrupted system call should be restarted Image flatten: 0% complete...2021-04-29 10:50:34.829 7ff7caffd700 -1 librbd::operation::FlattenRequest: 0x7ff7c445b470 should_complete: encountered error: (85) Interrupted system call should be restarted Image flatten: 39% complete...2021-04-29 10:50:42.081 7ff7caffd700 -1 librbd::operation::FlattenRequest: 0x7ff7c40324e0 should_complete: encountered error: (85) Interrupted system call should be restarted Image flatten: 0% complete...2021-04-29 10:50:43.897 7ff7caffd700 -1 librbd::operation::FlattenRequest: 0x7ff7c4018890 should_complete: encountered error: (85) Interrupted system call should be restarted Image flatten: 42% complete...2021-04-29 10:51:07.813 7ff7caffd700 -1 librbd::operation::FlattenRequest: 0x7ff7c402fe80 should_complete: encountered error: (85) Interrupted system call should be restarted Image flatten: 42% complete...2021-04-29 10:51:29.372 7ff7caffd700 -1 librbd::operation::FlattenRequest: 0x7ff7c40017c0 should_complete: encountered error: (85) Interrupted system call should be restarted


[email protected]:~# uname -a
Linux sm-node1.in.illusion.hu 5.4.106-1-pve #1 SMP PVE 5.4.106-1 (Fri, 19 Mar 2021 11:08:47 +0100) x86_64 GNU/Linux

[email protected]:~# dpkg -l |grep ceph
ii  ceph 14.2.20-pve1                            amd64        distributed storage and file system ii  ceph-base 14.2.20-pve1                            amd64        common ceph daemon libraries and management tools ii  ceph-common 14.2.20-pve1                            amd64        common utilities to mount and interact with a ceph storage cluster ii  ceph-fuse 14.2.20-pve1                            amd64        FUSE-based client for the Ceph distributed file system ii  ceph-mds 14.2.20-pve1                            amd64        metadata server for the ceph distributed file system ii  ceph-mgr 14.2.20-pve1                            amd64        manager for the ceph distributed storage system ii  ceph-mgr-dashboard 14.2.20-pve1                            all          dashboard plugin for ceph-mgr ii  ceph-mon 14.2.20-pve1                            amd64        monitor server for the ceph storage system ii  ceph-osd 14.2.20-pve1                            amd64        OSD server for the ceph storage system ii  libcephfs2 14.2.20-pve1                            amd64        Ceph distributed file system client library ii  python-ceph-argparse 14.2.20-pve1                            all          Python 2 utility libraries for Ceph CLI ii  python-cephfs 14.2.20-pve1                            amd64        Python 2 libraries for the Ceph libcephfs library


[email protected]:~# dpkg -l |grep rbd
ii  librbd1 14.2.20-pve1                            amd64        RADOS block device client library ii  python-rbd 14.2.20-pve1                            amd64        Python 2 libraries for the Ceph librbd library

[email protected]:~# modinfo rbd
filename: /lib/modules/5.4.106-1-pve/kernel/drivers/block/rbd.ko
license:        GPL
description:    RADOS Block Device (RBD) driver
author:         Jeff Garzik <[email protected]>
author:         Yehuda Sadeh <[email protected]>
author:         Sage Weil <[email protected]>
author:         Alex Elder <[email protected]>
srcversion:     7BA6FEE20249E416B2D09AB
depends:        libceph
retpoline:      Y
intree:         Y
name:           rbd
vermagic:       5.4.106-1-pve SMP mod_unload modversions
parm:           single_major:Use a single major number for all rbd devices (default: true) (bool)

Please give me advice, how can I produce more information to catch this.

Thank you,

i.

_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to