Yes, I'm on 10.3 on a brand new installation (i.e.: no upgrade or whatsoever)
Ok, I've finally read how to get core dumps on Debian. Soft limits are 0 by default, so no core dumps can be generated. I've set up soft ulimit to unlimited and killed a test process with a SIGSEGV signal, then I was able to see the core dump. Great! I've moved all the qcow2 files to another location but not destroyed the volume or deleted any files, I will create some test VMs there and leave this a few days to see if the dump is generated properly next time, hopefully soon enough. Thank you Xavi. *Angel Docampo* <https://www.google.com/maps/place/Edificio+de+Oficinas+Euro+3/@41.3755943,2.0730134,17z/data=!3m2!4b1!5s0x12a4997021aad323:0x3e06bf8ae6d68351!4m5!3m4!1s0x12a4997a67bf592f:0x83c2323a9cc2aa4b!8m2!3d41.3755903!4d2.0752021> <angel.doca...@eoniantec.com> <+34-93-1592929> El jue, 1 dic 2022 a las 18:02, Xavi Hernandez (<jaher...@redhat.com>) escribió: > I'm not so sure the problem is with sharding. Basically it's saying that > seek is not supported, which means that something between shard and the > bricks doesn't support it. DHT didn't support seek before 10.3, but if I'm > not wrong you are already using 10.3, so the message is weird. But in any > case this shouldn't cause a crash. The stack trace seems to indicate that > the crash happens inside disperse, but without a core dump there's little > more that I can do. > > > > On Thu, Dec 1, 2022 at 5:27 PM Angel Docampo <angel.doca...@eoniantec.com> > wrote: > >> Well, that last more time, but it crashed once again, same node, same >> mountpoint... fortunately, I've moved preventively all the VMs to the >> underlying ZFS filesystem those past days, so none of them have been >> affected this time... >> >> dmesg show this >> [2022-12-01 15:49:54] INFO: task iou-wrk-637144:946532 blocked for more >> than 120 seconds. >> [2022-12-01 15:49:54] Tainted: P IO 5.15.74-1-pve #1 >> [2022-12-01 15:49:54] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" >> disables this message. >> [2022-12-01 15:49:54] task:iou-wrk-637144 state:D stack: 0 >> pid:946532 ppid: 1 flags:0x00004000 >> [2022-12-01 15:49:54] Call Trace: >> [2022-12-01 15:49:54] <TASK> >> [2022-12-01 15:49:54] __schedule+0x34e/0x1740 >> [2022-12-01 15:49:54] ? kmem_cache_free+0x271/0x290 >> [2022-12-01 15:49:54] ? mempool_free_slab+0x17/0x20 >> [2022-12-01 15:49:54] schedule+0x69/0x110 >> [2022-12-01 15:49:54] rwsem_down_write_slowpath+0x231/0x4f0 >> [2022-12-01 15:49:54] ? ttwu_queue_wakelist+0x40/0x1c0 >> [2022-12-01 15:49:54] down_write+0x47/0x60 >> [2022-12-01 15:49:54] fuse_file_write_iter+0x1a3/0x430 >> [2022-12-01 15:49:54] ? apparmor_file_permission+0x70/0x170 >> [2022-12-01 15:49:54] io_write+0xf6/0x330 >> [2022-12-01 15:49:54] ? update_cfs_group+0x9c/0xc0 >> [2022-12-01 15:49:54] ? dequeue_entity+0xd8/0x490 >> [2022-12-01 15:49:54] io_issue_sqe+0x401/0x1fc0 >> [2022-12-01 15:49:54] ? lock_timer_base+0x3b/0xd0 >> [2022-12-01 15:49:54] io_wq_submit_work+0x76/0xd0 >> [2022-12-01 15:49:54] io_worker_handle_work+0x1a7/0x5f0 >> [2022-12-01 15:49:54] io_wqe_worker+0x2c0/0x360 >> [2022-12-01 15:49:54] ? finish_task_switch.isra.0+0x7e/0x2b0 >> [2022-12-01 15:49:54] ? io_worker_handle_work+0x5f0/0x5f0 >> [2022-12-01 15:49:54] ? io_worker_handle_work+0x5f0/0x5f0 >> [2022-12-01 15:49:54] ret_from_fork+0x1f/0x30 >> [2022-12-01 15:49:54] RIP: 0033:0x0 >> [2022-12-01 15:49:54] RSP: 002b:0000000000000000 EFLAGS: 00000207 >> [2022-12-01 15:49:54] RAX: 0000000000000000 RBX: 0000000000000011 RCX: >> 0000000000000000 >> [2022-12-01 15:49:54] RDX: 0000000000000001 RSI: 0000000000000001 RDI: >> 0000000000000120 >> [2022-12-01 15:49:54] RBP: 0000000000000120 R08: 0000000000000001 R09: >> 00000000000000f0 >> [2022-12-01 15:49:54] R10: 00000000000000f8 R11: 00000001239a4128 R12: >> ffffffffffffdb90 >> [2022-12-01 15:49:54] R13: 0000000000000001 R14: 0000000000000001 R15: >> 0000000000000100 >> [2022-12-01 15:49:54] </TASK> >> >> My gluster volume log shows plenty of error like this >> The message "I [MSGID: 133017] [shard.c:7275:shard_seek] 0-vmdata-shard: >> seek called on 73f0ad95-f7e3-4a68-8d08-9f7e03182baa. [Operation not >> supported]" repeated 1564 times between [2022-12-01 00:20:09.578233 +0000] >> and [2022-12-01 00:22:09.436927 +0000] >> [2022-12-01 00:22:09.516269 +0000] I [MSGID: 133017] >> [shard.c:7275:shard_seek] 0-vmdata-shard: seek called on >> 73f0ad95-f7e3-4a68-8d08-9f7e03182baa. [Operation not supported] >> >> and of this >> [2022-12-01 09:05:08.525867 +0000] I [MSGID: 133017] >> [shard.c:7275:shard_seek] 0-vmdata-shard: seek called on >> 3ed993c4-bbb5-4938-86e9-6d22b8541e8e. [Operation not supported] >> >> Then simply the same >> pending frames: >> frame : type(0) op(0) >> frame : type(0) op(0) >> frame : type(0) op(0) >> frame : type(0) op(0) >> frame : type(1) op(FSYNC) >> frame : type(0) op(0) >> frame : type(0) op(0) >> frame : type(0) op(0) >> frame : type(0) op(0) >> frame : type(0) op(0) >> frame : type(0) op(0) >> frame : type(0) op(0) >> frame : type(0) op(0) >> frame : type(0) op(0) >> frame : type(0) op(0) >> frame : type(0) op(0) >> frame : type(0) op(0) >> frame : type(0) op(0) >> frame : type(0) op(0) >> patchset: git://git.gluster.org/glusterfs.git >> signal received: 11 >> time of crash: >> 2022-12-01 14:45:14 +0000 >> configuration details: >> argp 1 >> backtrace 1 >> dlfcn 1 >> libpthread 1 >> llistxattr 1 >> setfsid 1 >> epoll.h 1 >> xattr.h 1 >> st_atim.tv_nsec 1 >> package-string: glusterfs 10.3 >> /lib/x86_64-linux-gnu/libglusterfs.so.0(+0x28a54)[0x7f1e23db3a54] >> /lib/x86_64-linux-gnu/libglusterfs.so.0(gf_print_trace+0x700)[0x7f1e23dbbfc0] >> >> /lib/x86_64-linux-gnu/libc.so.6(+0x38d60)[0x7f1e23b76d60] >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x37a14)[0x7f1e200e9a14] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x19414)[0x7f1e200cb414] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0xd072)[0x7f1e200bf072] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/performance/readdir-ahead.so(+0x316d)[0x7f1e200a316d] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/distribute.so(+0x5bdd4)[0x7f1e197aadd4] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/features/shard.so(+0x1e69c)[0x7f1e2008b69c] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/features/shard.so(+0x16551)[0x7f1e20083551] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/features/shard.so(+0x25abf)[0x7f1e20092abf] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/features/shard.so(+0x25d21)[0x7f1e20092d21] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/features/shard.so(+0x167be)[0x7f1e200837be] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/features/shard.so(+0x1c178)[0x7f1e20089178] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/features/utime.so(+0x7804)[0x7f1e20064804] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/performance/write-behind.so(+0x8164)[0x7f1e2004e164] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/performance/write-behind.so(+0x9228)[0x7f1e2004f228] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/performance/write-behind.so(+0x9a4d)[0x7f1e2004fa4d] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/features/utime.so(+0x29e5)[0x7f1e2005f9e5] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/features/shard.so(+0x12e59)[0x7f1e2007fe59] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/features/shard.so(+0xc2c6)[0x7f1e200792c6] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/distribute.so(+0x69e90)[0x7f1e197b8e90] >> >> /lib/x86_64-linux-gnu/libglusterfs.so.0(default_fxattrop_cbk+0x125)[0x7f1e23e27515] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x2421c)[0x7f1e200d621c] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x19414)[0x7f1e200cb414] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x16373)[0x7f1e200c8373] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x170f9)[0x7f1e200c90f9] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x1f929)[0x7f1e200d1929] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/protocol/client.so(+0x469c2)[0x7f1e201859c2] >> >> /lib/x86_64-linux-gnu/libgfrpc.so.0(+0xfccb)[0x7f1e23d5eccb] >> /lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_transport_notify+0x26)[0x7f1e23d5a646] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/rpc-transport/socket.so(+0x64c8)[0x7f1e202784c8] >> >> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/rpc-transport/socket.so(+0xd38c)[0x7f1e2027f38c] >> >> /lib/x86_64-linux-gnu/libglusterfs.so.0(+0x7971d)[0x7f1e23e0471d] >> /lib/x86_64-linux-gnu/libpthread.so.0(+0x7ea7)[0x7f1e23d1aea7] >> /lib/x86_64-linux-gnu/libc.so.6(clone+0x3f)[0x7f1e23c3aa2f] >> >> I'm still unable to gather any core dump. >> >> I can barely read something intelligible from all of this, but it's >> clearly something happening with sharding here. So I'm going to empty all >> the volume and destroy it completely and re-create another volume without >> sharding, and see what happens >> >> *Angel Docampo* >> >> <https://www.google.com/maps/place/Edificio+de+Oficinas+Euro+3/@41.3755943,2.0730134,17z/data=!3m2!4b1!5s0x12a4997021aad323:0x3e06bf8ae6d68351!4m5!3m4!1s0x12a4997a67bf592f:0x83c2323a9cc2aa4b!8m2!3d41.3755903!4d2.0752021> >> <angel.doca...@eoniantec.com> <+34-93-1592929> >> >> >> El vie, 25 nov 2022 a las 19:08, Angel Docampo (< >> angel.doca...@eoniantec.com>) escribió: >> >>> I did also notice about that loop0... AFAIK, I wasn't using any loop >>> device, at least consciously. >>> After looking for the same messages at the other gluster/proxmox nodes, >>> I saw no trace of it. >>> Then I saw on that node, there is a single LXC container, which disk is >>> living on the glusterfs, and effectively, is using ext4. >>> After the crash of today, I was unable to boot it up again, and the logs >>> became silent, I did just try to boot it up, and immediately appeared this >>> on dmesg >>> [2022-11-25 18:04:18] loop0: detected capacity change from 0 to >>> 16777216 >>> [2022-11-25 18:04:18] EXT4-fs (loop0): error loading journal >>> [2022-11-25 18:05:26] loop0: detected capacity change from 0 to >>> 16777216 >>> [2022-11-25 18:05:26] EXT4-fs (loop0): INFO: recovery required on >>> readonly filesystem >>> [2022-11-25 18:05:26] EXT4-fs (loop0): write access unavailable, cannot >>> proceed (try mounting with noload) >>> >>> And the LXC container didn't boot up. I've manually moved the LXC >>> container to the underlying ZFS where gluster lives, and the LXC booted up >>> and the dmesg log shows >>> [2022-11-25 18:24:06] loop0: detected capacity change from 0 to >>> 16777216 >>> [2022-11-25 18:24:06] EXT4-fs warning (device loop0): >>> ext4_multi_mount_protect:326: MMP interval 42 higher than expected, please >>> wait. >>> [2022-11-25 18:24:50] EXT4-fs (loop0): mounted filesystem with ordered >>> data mode. Opts: (null). Quota mode: none. >>> >>> So, to recapitulate: >>> - the loop device on the host relies on the LXC, is not surprising, but >>> I didn't know it. >>> - the LXC container had a lot of I/O issues just before the two crashes, >>> the crash from today, and the crash 4 days ago, this Monday >>> - as side note, this gluster is in production since last Thursday, so >>> the first crash was exactly 4 days since this LXC was started with the >>> storage on the gluster, and exactly 4 days after, it crashed again. >>> - this crashes began to happen since the upgrade to gluster 10.3, >>> because it was working just fine with former versions of gluster (from 3.X >>> to 9.X), and from proxmox 5.X to proxmox 7.1, when all the issues begun, >>> now I'm on proxmox 7.2. >>> - underlying ZFS where gluster is, has no ZIL or ZLOG (it had before the >>> upgrade to gluster 10.3, but as I had to re-create the gluster, I decided >>> not to add them because all my disks are SSD, so there is no need to add >>> any of those), I've added them to test if the LXC container caused the same >>> issues, it did, so they don't seem to make any difference. >>> - there are more loop0 I/O errors on the dmesg besides the days of the >>> crash, but there are just "one" error per day, and not all days, but the >>> days gluster mountpoint become inaccessible, there are tens of errors per >>> millisecond just before the crash >>> >>> I'm going to get rid of that LXC, as now I'm migrating from VMs to K8s >>> (living in a VM cluster inside proxmox), I was ready to convert this as >>> well, now is a must. >>> >>> I don't know if anyone at gluster can replicate this scenario (proxmox + >>> gluster distributed disperse + LXC on a gluster directory), to see if it >>> can be reproducible. I know this must be a corner case, just wondering why >>> stopped working, if it is a bug on GlusterFS 10.3, a bug in LXC or in >>> Proxmox 7.1 upwards (where I'm going to post this now, but Proxmox probably >>> won't be interested as they explicitly suggest mounting glusterfs with the >>> gluster client, and not to map a directory where gluster is mounted via >>> fstab) >>> >>> Thank you a lot Xavi, I will monitor dmesg to make sure all those loop >>> errors disappear, and hopefully I won't have a crash next Tuesday. :) >>> >>> *Angel Docampo* >>> >>> <https://www.google.com/maps/place/Edificio+de+Oficinas+Euro+3/@41.3755943,2.0730134,17z/data=!3m2!4b1!5s0x12a4997021aad323:0x3e06bf8ae6d68351!4m5!3m4!1s0x12a4997a67bf592f:0x83c2323a9cc2aa4b!8m2!3d41.3755903!4d2.0752021> >>> <angel.doca...@eoniantec.com> <+34-93-1592929> >>> >>> >>> El vie, 25 nov 2022 a las 13:25, Xavi Hernandez (<jaher...@redhat.com>) >>> escribió: >>> >>>> What is "loop0" it seems it's having some issue. Does it point to a >>>> Gluster file ? >>>> >>>> I also see that there's an io_uring thread in D state. If that one >>>> belongs to Gluster, it may explain why systemd was unable to generate a >>>> core dump (all threads need to be stopped to generate a core dump, but a >>>> thread blocked inside the kernel cannot be stopped). >>>> >>>> If you are using io_uring in Gluster, maybe you can disable it to see >>>> if it's related. >>>> >>>> Xavi >>>> >>>> On Fri, Nov 25, 2022 at 11:39 AM Angel Docampo < >>>> angel.doca...@eoniantec.com> wrote: >>>> >>>>> Well, just happened again, the same server, the same mountpoint. >>>>> >>>>> I'm unable to get the core dumps, coredumpctl says there are no core >>>>> dumps, it would be funny if I wasn't the one suffering it, but >>>>> systemd-coredump service crashed as well >>>>> ● systemd-coredump@0-3199871-0.service - Process Core Dump (PID >>>>> 3199871/UID 0) >>>>> Loaded: loaded (/lib/systemd/system/systemd-coredump@.service; >>>>> static) >>>>> Active: failed (Result: timeout) since Fri 2022-11-25 10:54:59 >>>>> CET; 39min ago >>>>> TriggeredBy: ● systemd-coredump.socket >>>>> Docs: man:systemd-coredump(8) >>>>> Process: 3199873 ExecStart=/lib/systemd/systemd-coredump >>>>> (code=killed, signal=TERM) >>>>> Main PID: 3199873 (code=killed, signal=TERM) >>>>> CPU: 15ms >>>>> >>>>> Nov 25 10:49:59 pve02 systemd[1]: Started Process Core Dump (PID >>>>> 3199871/UID 0). >>>>> Nov 25 10:54:59 pve02 systemd[1]: systemd-coredump@0-3199871-0.service: >>>>> Service reached runtime time limit. Stopping. >>>>> Nov 25 10:54:59 pve02 systemd[1]: systemd-coredump@0-3199871-0.service: >>>>> Failed with result 'timeout'. >>>>> >>>>> >>>>> I just saw the exception on dmesg, >>>>> [2022-11-25 10:50:08] INFO: task kmmpd-loop0:681644 blocked for more >>>>> than 120 seconds. >>>>> [2022-11-25 10:50:08] Tainted: P IO 5.15.60-2-pve >>>>> #1 >>>>> [2022-11-25 10:50:08] "echo 0 > >>>>> /proc/sys/kernel/hung_task_timeout_secs" disables this message. >>>>> [2022-11-25 10:50:08] task:kmmpd-loop0 state:D stack: 0 >>>>> pid:681644 ppid: 2 flags:0x00004000 >>>>> [2022-11-25 10:50:08] Call Trace: >>>>> [2022-11-25 10:50:08] <TASK> >>>>> [2022-11-25 10:50:08] __schedule+0x33d/0x1750 >>>>> [2022-11-25 10:50:08] ? bit_wait+0x70/0x70 >>>>> [2022-11-25 10:50:08] schedule+0x4e/0xc0 >>>>> [2022-11-25 10:50:08] io_schedule+0x46/0x80 >>>>> [2022-11-25 10:50:08] bit_wait_io+0x11/0x70 >>>>> [2022-11-25 10:50:08] __wait_on_bit+0x31/0xa0 >>>>> [2022-11-25 10:50:08] out_of_line_wait_on_bit+0x8d/0xb0 >>>>> [2022-11-25 10:50:08] ? var_wake_function+0x30/0x30 >>>>> [2022-11-25 10:50:08] __wait_on_buffer+0x34/0x40 >>>>> [2022-11-25 10:50:08] write_mmp_block+0x127/0x180 >>>>> [2022-11-25 10:50:08] kmmpd+0x1b9/0x430 >>>>> [2022-11-25 10:50:08] ? write_mmp_block+0x180/0x180 >>>>> [2022-11-25 10:50:08] kthread+0x127/0x150 >>>>> [2022-11-25 10:50:08] ? set_kthread_struct+0x50/0x50 >>>>> [2022-11-25 10:50:08] ret_from_fork+0x1f/0x30 >>>>> [2022-11-25 10:50:08] </TASK> >>>>> [2022-11-25 10:50:08] INFO: task iou-wrk-1511979:3200401 blocked for >>>>> more than 120 seconds. >>>>> [2022-11-25 10:50:08] Tainted: P IO 5.15.60-2-pve >>>>> #1 >>>>> [2022-11-25 10:50:08] "echo 0 > >>>>> /proc/sys/kernel/hung_task_timeout_secs" disables this message. >>>>> [2022-11-25 10:50:08] task:iou-wrk-1511979 state:D stack: 0 >>>>> pid:3200401 ppid: 1 flags:0x00004000 >>>>> [2022-11-25 10:50:08] Call Trace: >>>>> [2022-11-25 10:50:08] <TASK> >>>>> [2022-11-25 10:50:08] __schedule+0x33d/0x1750 >>>>> [2022-11-25 10:50:08] schedule+0x4e/0xc0 >>>>> [2022-11-25 10:50:08] rwsem_down_write_slowpath+0x231/0x4f0 >>>>> [2022-11-25 10:50:08] down_write+0x47/0x60 >>>>> [2022-11-25 10:50:08] fuse_file_write_iter+0x1a3/0x430 >>>>> [2022-11-25 10:50:08] ? apparmor_file_permission+0x70/0x170 >>>>> [2022-11-25 10:50:08] io_write+0xfb/0x320 >>>>> [2022-11-25 10:50:08] ? put_dec+0x1c/0xa0 >>>>> [2022-11-25 10:50:08] io_issue_sqe+0x401/0x1fc0 >>>>> [2022-11-25 10:50:08] io_wq_submit_work+0x76/0xd0 >>>>> [2022-11-25 10:50:08] io_worker_handle_work+0x1a7/0x5f0 >>>>> [2022-11-25 10:50:08] io_wqe_worker+0x2c0/0x360 >>>>> [2022-11-25 10:50:08] ? finish_task_switch.isra.0+0x7e/0x2b0 >>>>> [2022-11-25 10:50:08] ? io_worker_handle_work+0x5f0/0x5f0 >>>>> [2022-11-25 10:50:08] ? io_worker_handle_work+0x5f0/0x5f0 >>>>> [2022-11-25 10:50:08] ret_from_fork+0x1f/0x30 >>>>> [2022-11-25 10:50:08] RIP: 0033:0x0 >>>>> [2022-11-25 10:50:08] RSP: 002b:0000000000000000 EFLAGS: 00000216 >>>>> ORIG_RAX: 00000000000001aa >>>>> [2022-11-25 10:50:08] RAX: 0000000000000000 RBX: 00007fdb1efef640 >>>>> RCX: 00007fdd59f872e9 >>>>> [2022-11-25 10:50:08] RDX: 0000000000000000 RSI: 0000000000000001 >>>>> RDI: 0000000000000011 >>>>> [2022-11-25 10:50:08] RBP: 0000000000000000 R08: 0000000000000000 >>>>> R09: 0000000000000008 >>>>> [2022-11-25 10:50:08] R10: 0000000000000000 R11: 0000000000000216 >>>>> R12: 000055662e5bd268 >>>>> [2022-11-25 10:50:08] R13: 000055662e5bd320 R14: 000055662e5bd260 >>>>> R15: 0000000000000000 >>>>> [2022-11-25 10:50:08] </TASK> >>>>> [2022-11-25 10:52:08] INFO: task kmmpd-loop0:681644 blocked for more >>>>> than 241 seconds. >>>>> [2022-11-25 10:52:08] Tainted: P IO 5.15.60-2-pve >>>>> #1 >>>>> [2022-11-25 10:52:08] "echo 0 > >>>>> /proc/sys/kernel/hung_task_timeout_secs" disables this message. >>>>> [2022-11-25 10:52:08] task:kmmpd-loop0 state:D stack: 0 >>>>> pid:681644 ppid: 2 flags:0x00004000 >>>>> [2022-11-25 10:52:08] Call Trace: >>>>> [2022-11-25 10:52:08] <TASK> >>>>> [2022-11-25 10:52:08] __schedule+0x33d/0x1750 >>>>> [2022-11-25 10:52:08] ? bit_wait+0x70/0x70 >>>>> [2022-11-25 10:52:08] schedule+0x4e/0xc0 >>>>> [2022-11-25 10:52:08] io_schedule+0x46/0x80 >>>>> [2022-11-25 10:52:08] bit_wait_io+0x11/0x70 >>>>> [2022-11-25 10:52:08] __wait_on_bit+0x31/0xa0 >>>>> [2022-11-25 10:52:08] out_of_line_wait_on_bit+0x8d/0xb0 >>>>> [2022-11-25 10:52:08] ? var_wake_function+0x30/0x30 >>>>> [2022-11-25 10:52:08] __wait_on_buffer+0x34/0x40 >>>>> [2022-11-25 10:52:08] write_mmp_block+0x127/0x180 >>>>> [2022-11-25 10:52:08] kmmpd+0x1b9/0x430 >>>>> [2022-11-25 10:52:08] ? write_mmp_block+0x180/0x180 >>>>> [2022-11-25 10:52:08] kthread+0x127/0x150 >>>>> [2022-11-25 10:52:08] ? set_kthread_struct+0x50/0x50 >>>>> [2022-11-25 10:52:08] ret_from_fork+0x1f/0x30 >>>>> [2022-11-25 10:52:08] </TASK> >>>>> [2022-11-25 10:52:08] INFO: task iou-wrk-1511979:3200401 blocked for >>>>> more than 241 seconds. >>>>> [2022-11-25 10:52:08] Tainted: P IO 5.15.60-2-pve >>>>> #1 >>>>> [2022-11-25 10:52:08] "echo 0 > >>>>> /proc/sys/kernel/hung_task_timeout_secs" disables this message. >>>>> [2022-11-25 10:52:08] task:iou-wrk-1511979 state:D stack: 0 >>>>> pid:3200401 ppid: 1 flags:0x00004000 >>>>> [2022-11-25 10:52:08] Call Trace: >>>>> [2022-11-25 10:52:08] <TASK> >>>>> [2022-11-25 10:52:08] __schedule+0x33d/0x1750 >>>>> [2022-11-25 10:52:08] schedule+0x4e/0xc0 >>>>> [2022-11-25 10:52:08] rwsem_down_write_slowpath+0x231/0x4f0 >>>>> [2022-11-25 10:52:08] down_write+0x47/0x60 >>>>> [2022-11-25 10:52:08] fuse_file_write_iter+0x1a3/0x430 >>>>> [2022-11-25 10:52:08] ? apparmor_file_permission+0x70/0x170 >>>>> [2022-11-25 10:52:08] io_write+0xfb/0x320 >>>>> [2022-11-25 10:52:08] ? put_dec+0x1c/0xa0 >>>>> [2022-11-25 10:52:08] io_issue_sqe+0x401/0x1fc0 >>>>> [2022-11-25 10:52:08] io_wq_submit_work+0x76/0xd0 >>>>> [2022-11-25 10:52:08] io_worker_handle_work+0x1a7/0x5f0 >>>>> [2022-11-25 10:52:08] io_wqe_worker+0x2c0/0x360 >>>>> [2022-11-25 10:52:08] ? finish_task_switch.isra.0+0x7e/0x2b0 >>>>> [2022-11-25 10:52:08] ? io_worker_handle_work+0x5f0/0x5f0 >>>>> [2022-11-25 10:52:08] ? io_worker_handle_work+0x5f0/0x5f0 >>>>> [2022-11-25 10:52:08] ret_from_fork+0x1f/0x30 >>>>> [2022-11-25 10:52:08] RIP: 0033:0x0 >>>>> [2022-11-25 10:52:08] RSP: 002b:0000000000000000 EFLAGS: 00000216 >>>>> ORIG_RAX: 00000000000001aa >>>>> [2022-11-25 10:52:08] RAX: 0000000000000000 RBX: 00007fdb1efef640 >>>>> RCX: 00007fdd59f872e9 >>>>> [2022-11-25 10:52:08] RDX: 0000000000000000 RSI: 0000000000000001 >>>>> RDI: 0000000000000011 >>>>> [2022-11-25 10:52:08] RBP: 0000000000000000 R08: 0000000000000000 >>>>> R09: 0000000000000008 >>>>> [2022-11-25 10:52:08] R10: 0000000000000000 R11: 0000000000000216 >>>>> R12: 000055662e5bd268 >>>>> [2022-11-25 10:52:08] R13: 000055662e5bd320 R14: 000055662e5bd260 >>>>> R15: 0000000000000000 >>>>> [2022-11-25 10:52:08] </TASK> >>>>> [2022-11-25 10:52:12] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:52:12] print_req_error: 7 callbacks suppressed >>>>> [2022-11-25 10:52:12] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:52:12] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:52:12] EXT4-fs error (device loop0): kmmpd:179: comm >>>>> kmmpd-loop0: Error writing to MMP block >>>>> [2022-11-25 10:52:12] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:52:12] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:52:12] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:52:18] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:52:18] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:52:18] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:52:18] loop: Write error at byte offset 4490452992, >>>>> length 4096. >>>>> [2022-11-25 10:52:18] loop: Write error at byte offset 4490457088, >>>>> length 4096. >>>>> [2022-11-25 10:52:18] blk_update_request: I/O error, dev loop0, >>>>> sector 8770416 op 0x1:(WRITE) flags 0x800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:52:18] blk_update_request: I/O error, dev loop0, >>>>> sector 8770424 op 0x1:(WRITE) flags 0x800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:52:18] Aborting journal on device loop0-8. >>>>> [2022-11-25 10:52:18] loop: Write error at byte offset 4429185024, >>>>> length 4096. >>>>> [2022-11-25 10:52:18] blk_update_request: I/O error, dev loop0, >>>>> sector 8650752 op 0x1:(WRITE) flags 0x800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:52:18] blk_update_request: I/O error, dev loop0, >>>>> sector 8650752 op 0x1:(WRITE) flags 0x800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:52:18] Buffer I/O error on dev loop0, logical block >>>>> 1081344, lost sync page write >>>>> [2022-11-25 10:52:18] JBD2: Error -5 detected when updating journal >>>>> superblock for loop0-8. >>>>> [2022-11-25 10:52:23] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:52:23] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:52:23] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:52:28] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:52:28] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:52:28] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:52:33] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:52:33] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:52:33] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:52:38] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:52:38] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:52:38] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:52:43] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:52:43] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:52:43] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:52:48] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:52:48] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:52:48] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:52:53] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:52:53] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:52:53] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:52:59] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:52:59] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:52:59] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:53:04] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:53:04] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:53:04] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:53:09] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:53:09] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:53:09] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:53:14] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:53:14] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:53:14] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:53:19] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:53:19] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:53:19] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:53:24] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:53:24] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:53:24] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:53:29] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:53:29] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:53:29] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:53:34] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:53:34] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:53:34] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:53:40] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:53:40] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:53:40] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:53:45] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:53:45] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:53:45] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:53:50] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:53:50] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:53:50] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:53:55] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:53:55] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:53:55] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:54:00] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:54:00] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:54:00] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:54:05] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:54:05] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:54:05] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:54:10] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:54:10] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:54:10] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:54:15] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:54:15] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:54:15] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:54:21] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:54:21] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:54:21] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:54:26] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:54:26] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:54:26] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:54:31] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:54:31] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:54:31] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:54:36] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:54:36] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:54:36] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:54:41] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:54:41] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:54:41] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:54:46] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:54:46] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:54:46] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:54:51] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:54:51] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:54:51] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:54:56] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:54:56] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:54:56] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:55:01] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:55:01] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:55:01] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:55:04] EXT4-fs error (device loop0): >>>>> ext4_journal_check_start:83: comm burp: Detected aborted journal >>>>> [2022-11-25 10:55:04] loop: Write error at byte offset 0, length >>>>> 4096. >>>>> [2022-11-25 10:55:04] blk_update_request: I/O error, dev loop0, >>>>> sector 0 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:55:04] blk_update_request: I/O error, dev loop0, >>>>> sector 0 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:55:04] Buffer I/O error on dev loop0, logical block 0, >>>>> lost sync page write >>>>> [2022-11-25 10:55:04] EXT4-fs (loop0): I/O error while writing >>>>> superblock >>>>> [2022-11-25 10:55:04] EXT4-fs (loop0): Remounting filesystem >>>>> read-only >>>>> [2022-11-25 10:55:07] loop: Write error at byte offset 37908480, >>>>> length 4096. >>>>> [2022-11-25 10:55:07] blk_update_request: I/O error, dev loop0, >>>>> sector 74040 op 0x1:(WRITE) flags 0x3800 phys_seg 1 prio class 0 >>>>> [2022-11-25 10:55:07] Buffer I/O error on dev loop0, logical block >>>>> 9255, lost sync page write >>>>> [2022-11-25 10:57:14] blk_update_request: I/O error, dev loop0, >>>>> sector 16390368 op 0x0:(READ) flags 0x80700 phys_seg 6 prio class 0 >>>>> [2022-11-25 11:03:45] device tap136i0 entered promiscuous mode >>>>> >>>>> I don't know if it is relevant somehow or it is unrelated to >>>>> glusterfs, but the consequences are the mountpoint crashes, I'm forced to >>>>> lazy unmount it and remount it back. Then restart all the VMs on there, >>>>> unfortunately, this time several have the hard disk corrupted and now I'm >>>>> restoring them from the backup. >>>>> >>>>> Any tip? >>>>> >>>>> *Angel Docampo* >>>>> >>>>> <https://www.google.com/maps/place/Edificio+de+Oficinas+Euro+3/@41.3755943,2.0730134,17z/data=!3m2!4b1!5s0x12a4997021aad323:0x3e06bf8ae6d68351!4m5!3m4!1s0x12a4997a67bf592f:0x83c2323a9cc2aa4b!8m2!3d41.3755903!4d2.0752021> >>>>> <angel.doca...@eoniantec.com> <+34-93-1592929> >>>>> >>>>> >>>>> El mar, 22 nov 2022 a las 12:31, Angel Docampo (< >>>>> angel.doca...@eoniantec.com>) escribió: >>>>> >>>>>> I've taken a look into all possible places they should be, and I >>>>>> couldn't find it anywhere. Some people say the dump file is generated >>>>>> where >>>>>> the application is running... well, I don't know where to look then, and >>>>>> I >>>>>> hope they hadn't been generated on the failed mountpoint. >>>>>> >>>>>> As Debian 11 has systemd, I've installed systemd-coredump, so in the >>>>>> case a new crash happens, at least I will have the exact location and >>>>>> tool >>>>>> (coredumpctl) to find them and will install then the debug symbols, which >>>>>> is particularly tricky on debian. But I need to wait to happen again, now >>>>>> the tool says there isn't any core dump on the system. >>>>>> >>>>>> Thank you, Xavi, if this happens again (let's hope it won't), I will >>>>>> report back. >>>>>> >>>>>> Best regards! >>>>>> >>>>>> *Angel Docampo* >>>>>> >>>>>> <https://www.google.com/maps/place/Edificio+de+Oficinas+Euro+3/@41.3755943,2.0730134,17z/data=!3m2!4b1!5s0x12a4997021aad323:0x3e06bf8ae6d68351!4m5!3m4!1s0x12a4997a67bf592f:0x83c2323a9cc2aa4b!8m2!3d41.3755903!4d2.0752021> >>>>>> <angel.doca...@eoniantec.com> <+34-93-1592929> >>>>>> >>>>>> >>>>>> El mar, 22 nov 2022 a las 10:45, Xavi Hernandez (<jaher...@redhat.com>) >>>>>> escribió: >>>>>> >>>>>>> The crash seems related to some problem in ec xlator, but I don't >>>>>>> have enough information to determine what it is. The crash should have >>>>>>> generated a core dump somewhere in the system (I don't know where Debian >>>>>>> keeps the core dumps). If you find it, you should be able to open it >>>>>>> using >>>>>>> this command (make sure debug symbols package is also installed before >>>>>>> running it): >>>>>>> >>>>>>> # gdb /usr/sbin/glusterfs <path to core dump> >>>>>>> >>>>>>> And then run this command: >>>>>>> >>>>>>> # bt -full >>>>>>> >>>>>>> Regards, >>>>>>> >>>>>>> Xavi >>>>>>> >>>>>>> On Tue, Nov 22, 2022 at 9:41 AM Angel Docampo < >>>>>>> angel.doca...@eoniantec.com> wrote: >>>>>>> >>>>>>>> Hi Xavi, >>>>>>>> >>>>>>>> The OS is Debian 11 with the proxmox kernel. Gluster packages are >>>>>>>> the official from gluster.org ( >>>>>>>> https://download.gluster.org/pub/gluster/glusterfs/10/10.3/Debian/bullseye/ >>>>>>>> ) >>>>>>>> >>>>>>>> The system logs showed no other issues by the time of the crash, no >>>>>>>> OOM kill or whatsoever, and no other process was interacting with the >>>>>>>> gluster mountpoint besides proxmox. >>>>>>>> >>>>>>>> I wasn't running gdb when it crashed, so I don't really know if I >>>>>>>> can obtain a more detailed trace from logs or if there is a simple way >>>>>>>> to >>>>>>>> let it running in the background to see if it happens again (or there >>>>>>>> is a >>>>>>>> flag to start the systemd daemon in debug mode). >>>>>>>> >>>>>>>> Best, >>>>>>>> >>>>>>>> *Angel Docampo* >>>>>>>> >>>>>>>> <https://www.google.com/maps/place/Edificio+de+Oficinas+Euro+3/@41.3755943,2.0730134,17z/data=!3m2!4b1!5s0x12a4997021aad323:0x3e06bf8ae6d68351!4m5!3m4!1s0x12a4997a67bf592f:0x83c2323a9cc2aa4b!8m2!3d41.3755903!4d2.0752021> >>>>>>>> <angel.doca...@eoniantec.com> <+34-93-1592929> >>>>>>>> >>>>>>>> >>>>>>>> El lun, 21 nov 2022 a las 15:16, Xavi Hernandez (< >>>>>>>> jaher...@redhat.com>) escribió: >>>>>>>> >>>>>>>>> Hi Angel, >>>>>>>>> >>>>>>>>> On Mon, Nov 21, 2022 at 2:33 PM Angel Docampo < >>>>>>>>> angel.doca...@eoniantec.com> wrote: >>>>>>>>> >>>>>>>>>> Sorry for necrobumping this, but this morning I've suffered this >>>>>>>>>> on my Proxmox + GlusterFS cluster. In the log I can see this >>>>>>>>>> >>>>>>>>>> [2022-11-21 07:38:00.213620 +0000] I [MSGID: 133017] >>>>>>>>>> [shard.c:7275:shard_seek] 11-vmdata-shard: seek called on >>>>>>>>>> fbc063cb-874e-475d-b585-f89 >>>>>>>>>> f7518acdd. [Operation not supported] >>>>>>>>>> pending frames: >>>>>>>>>> frame : type(1) op(WRITE) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> frame : type(0) op(0) >>>>>>>>>> ... >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> frame : type(1) op(FSYNC) >>>>>>>>>> patchset: git://git.gluster.org/glusterfs.git >>>>>>>>>> signal received: 11 >>>>>>>>>> time of crash: >>>>>>>>>> 2022-11-21 07:38:00 +0000 >>>>>>>>>> configuration details: >>>>>>>>>> argp 1 >>>>>>>>>> backtrace 1 >>>>>>>>>> dlfcn 1 >>>>>>>>>> libpthread 1 >>>>>>>>>> llistxattr 1 >>>>>>>>>> setfsid 1 >>>>>>>>>> epoll.h 1 >>>>>>>>>> xattr.h 1 >>>>>>>>>> st_atim.tv_nsec 1 >>>>>>>>>> package-string: glusterfs 10.3 >>>>>>>>>> /lib/x86_64-linux-gnu/libglusterfs.so.0(+0x28a54)[0x7f74f286ba54] >>>>>>>>>> /lib/x86_64-linux-gnu/libglusterfs.so.0(gf_print_trace+0x700)[0x7f74f2873fc0] >>>>>>>>>> >>>>>>>>>> /lib/x86_64-linux-gnu/libc.so.6(+0x38d60)[0x7f74f262ed60] >>>>>>>>>> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x37a14)[0x7f74ecfcea14] >>>>>>>>>> >>>>>>>>>> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x19414)[0x7f74ecfb0414] >>>>>>>>>> >>>>>>>>>> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x16373)[0x7f74ecfad373] >>>>>>>>>> >>>>>>>>>> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x21d59)[0x7f74ecfb8d59] >>>>>>>>>> >>>>>>>>>> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x22815)[0x7f74ecfb9815] >>>>>>>>>> >>>>>>>>>> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x377d9)[0x7f74ecfce7d9] >>>>>>>>>> >>>>>>>>>> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x19414)[0x7f74ecfb0414] >>>>>>>>>> >>>>>>>>>> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x16373)[0x7f74ecfad373] >>>>>>>>>> >>>>>>>>>> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x170f9)[0x7f74ecfae0f9] >>>>>>>>>> >>>>>>>>>> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/cluster/disperse.so(+0x313bb)[0x7f74ecfc83bb] >>>>>>>>>> >>>>>>>>>> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/xlator/protocol/client.so(+0x48e3a)[0x7f74ed06ce3a] >>>>>>>>>> >>>>>>>>>> /lib/x86_64-linux-gnu/libgfrpc.so.0(+0xfccb)[0x7f74f2816ccb] >>>>>>>>>> /lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_transport_notify+0x26)[0x7f74f2812646] >>>>>>>>>> >>>>>>>>>> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/rpc-transport/socket.so(+0x64c8)[0x7f74ee15f4c8] >>>>>>>>>> >>>>>>>>>> /usr/lib/x86_64-linux-gnu/glusterfs/10.3/rpc-transport/socket.so(+0xd38c)[0x7f74ee16638c] >>>>>>>>>> >>>>>>>>>> /lib/x86_64-linux-gnu/libglusterfs.so.0(+0x7971d)[0x7f74f28bc71d] >>>>>>>>>> /lib/x86_64-linux-gnu/libpthread.so.0(+0x7ea7)[0x7f74f27d2ea7] >>>>>>>>>> /lib/x86_64-linux-gnu/libc.so.6(clone+0x3f)[0x7f74f26f2aef] >>>>>>>>>> --------- >>>>>>>>>> The mount point wasn't accessible with the "Transport endpoint >>>>>>>>>> is not connected" message and it was shown like this. >>>>>>>>>> d????????? ? ? ? ? ? vmdata >>>>>>>>>> >>>>>>>>>> I had to stop all the VMs on that proxmox node, then stop the >>>>>>>>>> gluster daemon to ummount de directory, and after starting the >>>>>>>>>> daemon and >>>>>>>>>> re-mounting, all was working again. >>>>>>>>>> >>>>>>>>>> My gluster volume info returns this >>>>>>>>>> >>>>>>>>>> Volume Name: vmdata >>>>>>>>>> Type: Distributed-Disperse >>>>>>>>>> Volume ID: cace5aa4-b13a-4750-8736-aa179c2485e1 >>>>>>>>>> Status: Started >>>>>>>>>> Snapshot Count: 0 >>>>>>>>>> Number of Bricks: 2 x (2 + 1) = 6 >>>>>>>>>> Transport-type: tcp >>>>>>>>>> Bricks: >>>>>>>>>> Brick1: g01:/data/brick1/brick >>>>>>>>>> Brick2: g02:/data/brick2/brick >>>>>>>>>> Brick3: g03:/data/brick1/brick >>>>>>>>>> Brick4: g01:/data/brick2/brick >>>>>>>>>> Brick5: g02:/data/brick1/brick >>>>>>>>>> Brick6: g03:/data/brick2/brick >>>>>>>>>> Options Reconfigured: >>>>>>>>>> nfs.disable: on >>>>>>>>>> transport.address-family: inet >>>>>>>>>> storage.fips-mode-rchecksum: on >>>>>>>>>> features.shard: enable >>>>>>>>>> features.shard-block-size: 256MB >>>>>>>>>> performance.read-ahead: off >>>>>>>>>> performance.quick-read: off >>>>>>>>>> performance.io-cache: off >>>>>>>>>> server.event-threads: 2 >>>>>>>>>> client.event-threads: 3 >>>>>>>>>> performance.client-io-threads: on >>>>>>>>>> performance.stat-prefetch: off >>>>>>>>>> dht.force-readdirp: off >>>>>>>>>> performance.force-readdirp: off >>>>>>>>>> network.remote-dio: on >>>>>>>>>> features.cache-invalidation: on >>>>>>>>>> performance.parallel-readdir: on >>>>>>>>>> performance.readdir-ahead: on >>>>>>>>>> >>>>>>>>>> Xavi, do you think the open-behind off setting can help somehow? >>>>>>>>>> I did try to understand what it does (with no luck), and if it could >>>>>>>>>> impact >>>>>>>>>> the performance of my VMs (I've the setup you know so well ;)) >>>>>>>>>> I would like to avoid more crashings like this, version 10.3 of >>>>>>>>>> gluster was working since two weeks ago, quite well until this >>>>>>>>>> morning. >>>>>>>>>> >>>>>>>>> >>>>>>>>> I don't think disabling open-behind will have any visible effect >>>>>>>>> on performance. Open-behind is only useful for small files when the >>>>>>>>> workload is mostly open + read + close, and quick-read is also enabled >>>>>>>>> (which is not your case). The only effect it will have is that the >>>>>>>>> latency >>>>>>>>> "saved" during open is "paid" on the next operation sent to the file, >>>>>>>>> so >>>>>>>>> the total overall latency should be the same. Additionally, VM >>>>>>>>> workload >>>>>>>>> doesn't open files frequently, so it shouldn't matter much in any >>>>>>>>> case. >>>>>>>>> >>>>>>>>> That said, I'm not sure if the problem is the same in your case. >>>>>>>>> Based on the stack of the crash, it seems an issue inside the disperse >>>>>>>>> module. >>>>>>>>> >>>>>>>>> What OS are you using ? are you using official packages ? if so, >>>>>>>>> which ones ? >>>>>>>>> >>>>>>>>> Is it possible to provide a backtrace from gdb ? >>>>>>>>> >>>>>>>>> Regards, >>>>>>>>> >>>>>>>>> Xavi >>>>>>>>> >>>>>>>>> >>>>>>>>>> *Angel Docampo* >>>>>>>>>> >>>>>>>>>> <https://www.google.com/maps/place/Edificio+de+Oficinas+Euro+3/@41.3755943,2.0730134,17z/data=!3m2!4b1!5s0x12a4997021aad323:0x3e06bf8ae6d68351!4m5!3m4!1s0x12a4997a67bf592f:0x83c2323a9cc2aa4b!8m2!3d41.3755903!4d2.0752021> >>>>>>>>>> <angel.doca...@eoniantec.com> <+34-93-1592929> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> El vie, 19 mar 2021 a las 2:10, David Cunningham (< >>>>>>>>>> dcunning...@voisonics.com>) escribió: >>>>>>>>>> >>>>>>>>>>> Hi Xavi, >>>>>>>>>>> >>>>>>>>>>> Thank you for that information. We'll look at upgrading it. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Fri, 12 Mar 2021 at 05:20, Xavi Hernandez < >>>>>>>>>>> jaher...@redhat.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hi David, >>>>>>>>>>>> >>>>>>>>>>>> with so little information it's hard to tell, but given that >>>>>>>>>>>> there are several OPEN and UNLINK operations, it could be related >>>>>>>>>>>> to an >>>>>>>>>>>> already fixed bug (in recent versions) in open-behind. >>>>>>>>>>>> >>>>>>>>>>>> You can try disabling open-behind with this command: >>>>>>>>>>>> >>>>>>>>>>>> # gluster volume set <volname> open-behind off >>>>>>>>>>>> >>>>>>>>>>>> But given the version you are using is very old and >>>>>>>>>>>> unmaintained, I would recommend you to upgrade to 8.x at least. >>>>>>>>>>>> >>>>>>>>>>>> Regards, >>>>>>>>>>>> >>>>>>>>>>>> Xavi >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Wed, Mar 10, 2021 at 5:10 AM David Cunningham < >>>>>>>>>>>> dcunning...@voisonics.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Hello, >>>>>>>>>>>>> >>>>>>>>>>>>> We have a GlusterFS 5.13 server which also mounts itself with >>>>>>>>>>>>> the native FUSE client. Recently the FUSE mount crashed and we >>>>>>>>>>>>> found the >>>>>>>>>>>>> following in the syslog. There isn't anything logged in >>>>>>>>>>>>> mnt-glusterfs.log >>>>>>>>>>>>> for that time. After killing all processes with a file handle >>>>>>>>>>>>> open on the >>>>>>>>>>>>> filesystem we were able to unmount and then remount the filesystem >>>>>>>>>>>>> successfully. >>>>>>>>>>>>> >>>>>>>>>>>>> Would anyone have advice on how to debug this crash? Thank you >>>>>>>>>>>>> in advance! >>>>>>>>>>>>> >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: pending frames: >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: frame : type(0) op(0) >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: frame : type(0) op(0) >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: frame : type(1) >>>>>>>>>>>>> op(UNLINK) >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: frame : type(1) >>>>>>>>>>>>> op(UNLINK) >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: frame : type(1) >>>>>>>>>>>>> op(OPEN) >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: message repeated >>>>>>>>>>>>> 3355 times: [ frame : type(1) op(OPEN)] >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: frame : type(1) >>>>>>>>>>>>> op(OPEN) >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: message repeated >>>>>>>>>>>>> 6965 times: [ frame : type(1) op(OPEN)] >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: frame : type(1) >>>>>>>>>>>>> op(OPEN) >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: message repeated >>>>>>>>>>>>> 4095 times: [ frame : type(1) op(OPEN)] >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: frame : type(0) op(0) >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: patchset: git:// >>>>>>>>>>>>> git.gluster.org/glusterfs.git >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: signal received: 11 >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: time of crash: >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: 2021-03-09 03:12:31 >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: configuration >>>>>>>>>>>>> details: >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: argp 1 >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: backtrace 1 >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: dlfcn 1 >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: libpthread 1 >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: llistxattr 1 >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: setfsid 1 >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: spinlock 1 >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: epoll.h 1 >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: xattr.h 1 >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: st_atim.tv_nsec 1 >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: package-string: >>>>>>>>>>>>> glusterfs 5.13 >>>>>>>>>>>>> Mar 9 05:12:31 voip1 mnt-glusterfs[2932]: --------- >>>>>>>>>>>>> ... >>>>>>>>>>>>> Mar 9 05:13:50 voip1 systemd[1]: >>>>>>>>>>>>> glusterfssharedstorage.service: Main process exited, code=killed, >>>>>>>>>>>>> status=11/SEGV >>>>>>>>>>>>> Mar 9 05:13:50 voip1 systemd[1]: >>>>>>>>>>>>> glusterfssharedstorage.service: Failed with result 'signal'. >>>>>>>>>>>>> ... >>>>>>>>>>>>> Mar 9 05:13:54 voip1 systemd[1]: >>>>>>>>>>>>> glusterfssharedstorage.service: Service hold-off time over, >>>>>>>>>>>>> scheduling >>>>>>>>>>>>> restart. >>>>>>>>>>>>> Mar 9 05:13:54 voip1 systemd[1]: >>>>>>>>>>>>> glusterfssharedstorage.service: Scheduled restart job, restart >>>>>>>>>>>>> counter is >>>>>>>>>>>>> at 2. >>>>>>>>>>>>> Mar 9 05:13:54 voip1 systemd[1]: Stopped Mount glusterfs >>>>>>>>>>>>> sharedstorage. >>>>>>>>>>>>> Mar 9 05:13:54 voip1 systemd[1]: Starting Mount glusterfs >>>>>>>>>>>>> sharedstorage... >>>>>>>>>>>>> Mar 9 05:13:54 voip1 mount-shared-storage.sh[20520]: ERROR: >>>>>>>>>>>>> Mount point does not exist >>>>>>>>>>>>> Mar 9 05:13:54 voip1 mount-shared-storage.sh[20520]: Please >>>>>>>>>>>>> specify a mount point >>>>>>>>>>>>> Mar 9 05:13:54 voip1 mount-shared-storage.sh[20520]: Usage: >>>>>>>>>>>>> Mar 9 05:13:54 voip1 mount-shared-storage.sh[20520]: man 8 >>>>>>>>>>>>> /sbin/mount.glusterfs >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> David Cunningham, Voisonics Limited >>>>>>>>>>>>> http://voisonics.com/ >>>>>>>>>>>>> USA: +1 213 221 1092 >>>>>>>>>>>>> New Zealand: +64 (0)28 2558 3782 >>>>>>>>>>>>> ________ >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Community Meeting Calendar: >>>>>>>>>>>>> >>>>>>>>>>>>> Schedule - >>>>>>>>>>>>> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >>>>>>>>>>>>> Bridge: https://meet.google.com/cpu-eiue-hvk >>>>>>>>>>>>> Gluster-users mailing list >>>>>>>>>>>>> Gluster-users@gluster.org >>>>>>>>>>>>> https://lists.gluster.org/mailman/listinfo/gluster-users >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> David Cunningham, Voisonics Limited >>>>>>>>>>> http://voisonics.com/ >>>>>>>>>>> USA: +1 213 221 1092 >>>>>>>>>>> New Zealand: +64 (0)28 2558 3782 >>>>>>>>>>> ________ >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Community Meeting Calendar: >>>>>>>>>>> >>>>>>>>>>> Schedule - >>>>>>>>>>> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >>>>>>>>>>> Bridge: https://meet.google.com/cpu-eiue-hvk >>>>>>>>>>> Gluster-users mailing list >>>>>>>>>>> Gluster-users@gluster.org >>>>>>>>>>> https://lists.gluster.org/mailman/listinfo/gluster-users >>>>>>>>>>> >>>>>>>>>>
________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://meet.google.com/cpu-eiue-hvk Gluster-users mailing list Gluster-users@gluster.org https://lists.gluster.org/mailman/listinfo/gluster-users