At the moment, it is responding like I would expect. I do know I have one failed drive on one brick (hardware failure, OS removed drive completely; the underlying /dev/sdb is gone). I have a new disk on order (overnight), but that is also one brick of one volume that is replica 3, so I would hope a complete failure like that would restore the system to operational capabilities.
Since having the gluster-volume-starting problems, I have performed a test in the engine volume with writing and removing a file and verifying its happening from all three hosts; that worked. The engine volume has all of its bricks, as well two other volumes; its only one volume that is shy one brick. --Jim On Tue, May 29, 2018 at 11:41 PM, Johan Bernhardsson <[email protected]> wrote: > Is storage working as it should? Does the gluster mount point respond as > it should? Can you write files to it? Does the physical drives say that > they are ok? Can you write (you shouldn't bypass gluster mount point but > you need to test the drives) to the physical drives? > > For me this sounds like broken or almost broken hardware or broken > underlying filesystems. > > If one of the drives malfunction and timeout, gluster will be slow and > timeout. It runs write in sync so the slowest node will slow down the whole > system. > > /Johan > > > On May 30, 2018 08:29:46 Jim Kusznir <[email protected]> wrote: > >> hosted-engine --deploy failed (would not come up on my existing gluster >> storage). However, I realized no changes were written to my existing >> storage. So, I went back to trying to get my old engine running. >> >> hosted-engine --vm-status is now taking a very long time (5+minutes) to >> return, and it returns stail information everywhere. I thought perhaps the >> lockspace is corrupt, so tried to clean that and metadata, but both are >> failing (--cleam-metadata has hung and I can't even ctrl-c out of it). >> >> How can I reinitialize all the lockspace/metadata safely? There is no >> engine or VMs running currently.... >> >> --Jim >> >> On Tue, May 29, 2018 at 9:33 PM, Jim Kusznir <[email protected]> wrote: >> >>> Well, things went from bad to very, very bad.... >>> >>> It appears that during one of the 2 minute lockups, the fencing agents >>> decided that another node in the cluster was down. As a result, 2 of the 3 >>> nodes were simultaneously reset with fencing agent reboot. After the nodes >>> came back up, the engine would not start. All running VMs (including VMs >>> on the 3rd node that was not rebooted) crashed. >>> >>> I've now been working for about 3 hours trying to get the engine to come >>> up. I don't know why it won't start. hosted-engine --vm-start says its >>> starting, but it doesn't start (virsh doesn't show any VMs running). I'm >>> currently running --deploy, as I had run out of options for anything else I >>> can come up with. I hope this will allow me to re-import all my existing >>> VMs and allow me to start them back up after everything comes back up. >>> >>> I do have an unverified geo-rep backup; I don't know if it is a good >>> backup (there were several prior messages to this list, but I didn't get >>> replies to my questions. It was running in what I believe to be "strange", >>> and the data directories are larger than their source). >>> >>> I'll see if my --deploy works, and if not, I'll be back with another >>> message/help request. >>> >>> When the dust settles and I'm at least minimally functional again, I >>> really want to understand why all these technologies designed to offer >>> redundancy conspired to reduce uptime and create failures where there >>> weren't any otherwise. I thought with hosted engine, 3 ovirt servers and >>> glusterfs with minimum replica 2+arb or replica 3 should have offered >>> strong resilience against server failure or disk failure, and should have >>> prevented / recovered from data corruption. Instead, all of the above >>> happened (once I get my cluster back up, I still have to try and recover my >>> webserver VM, which won't boot due to XFS corrupt journal issues created >>> during the gluster crashes). I think a lot of these issues were rooted >>> from the upgrade from 4.1 to 4.2. >>> >>> --Jim >>> >>> On Tue, May 29, 2018 at 6:25 PM, Jim Kusznir <[email protected]> >>> wrote: >>> >>>> I also finally found the following in my system log on one server: >>>> >>>> [10679.524491] INFO: task glusterclogro:14933 blocked for more than 120 >>>> seconds. >>>> [10679.525826] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" >>>> disables this message. >>>> [10679.527144] glusterclogro D ffff97209832bf40 0 14933 1 >>>> 0x00000080 >>>> [10679.527150] Call Trace: >>>> [10679.527161] [<ffffffffb9913f79>] schedule+0x29/0x70 >>>> [10679.527218] [<ffffffffc060e388>] _xfs_log_force_lsn+0x2e8/0x340 >>>> [xfs] >>>> [10679.527225] [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 >>>> [10679.527254] [<ffffffffc05eeb97>] xfs_file_fsync+0x107/0x1e0 [xfs] >>>> [10679.527260] [<ffffffffb944f0e7>] do_fsync+0x67/0xb0 >>>> [10679.527268] [<ffffffffb992076f>] ? system_call_after_swapgs+0xbc/ >>>> 0x160 >>>> [10679.527271] [<ffffffffb944f3d0>] SyS_fsync+0x10/0x20 >>>> [10679.527275] [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 >>>> [10679.527279] [<ffffffffb992077b>] ? system_call_after_swapgs+0xc8/ >>>> 0x160 >>>> [10679.527283] INFO: task glusterposixfsy:14941 blocked for more than >>>> 120 seconds. >>>> [10679.528608] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" >>>> disables this message. >>>> [10679.529956] glusterposixfsy D ffff972495f84f10 0 14941 1 >>>> 0x00000080 >>>> [10679.529961] Call Trace: >>>> [10679.529966] [<ffffffffb9913f79>] schedule+0x29/0x70 >>>> [10679.530003] [<ffffffffc060e388>] _xfs_log_force_lsn+0x2e8/0x340 >>>> [xfs] >>>> [10679.530008] [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 >>>> [10679.530038] [<ffffffffc05eeb97>] xfs_file_fsync+0x107/0x1e0 [xfs] >>>> [10679.530042] [<ffffffffb944f0e7>] do_fsync+0x67/0xb0 >>>> [10679.530046] [<ffffffffb992076f>] ? system_call_after_swapgs+0xbc/ >>>> 0x160 >>>> [10679.530050] [<ffffffffb944f3f3>] SyS_fdatasync+0x13/0x20 >>>> [10679.530054] [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 >>>> [10679.530058] [<ffffffffb992077b>] ? system_call_after_swapgs+0xc8/ >>>> 0x160 >>>> [10679.530062] INFO: task glusteriotwr13:15486 blocked for more than >>>> 120 seconds. >>>> [10679.531805] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" >>>> disables this message. >>>> [10679.533732] glusteriotwr13 D ffff9720a83f0000 0 15486 1 >>>> 0x00000080 >>>> [10679.533738] Call Trace: >>>> [10679.533747] [<ffffffffb9913f79>] schedule+0x29/0x70 >>>> [10679.533799] [<ffffffffc060e388>] _xfs_log_force_lsn+0x2e8/0x340 >>>> [xfs] >>>> [10679.533806] [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 >>>> [10679.533846] [<ffffffffc05eeb97>] xfs_file_fsync+0x107/0x1e0 [xfs] >>>> [10679.533852] [<ffffffffb944f0e7>] do_fsync+0x67/0xb0 >>>> [10679.533858] [<ffffffffb992076f>] ? system_call_after_swapgs+0xbc/ >>>> 0x160 >>>> [10679.533863] [<ffffffffb944f3f3>] SyS_fdatasync+0x13/0x20 >>>> [10679.533868] [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 >>>> [10679.533873] [<ffffffffb992077b>] ? system_call_after_swapgs+0xc8/ >>>> 0x160 >>>> [10919.512757] INFO: task glusterclogro:14933 blocked for more than 120 >>>> seconds. >>>> [10919.514714] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" >>>> disables this message. >>>> [10919.516663] glusterclogro D ffff97209832bf40 0 14933 1 >>>> 0x00000080 >>>> [10919.516677] Call Trace: >>>> [10919.516690] [<ffffffffb9913f79>] schedule+0x29/0x70 >>>> [10919.516696] [<ffffffffb99118e9>] schedule_timeout+0x239/0x2c0 >>>> [10919.516703] [<ffffffffb951cc04>] ? blk_finish_plug+0x14/0x40 >>>> [10919.516768] [<ffffffffc05e9224>] ? _xfs_buf_ioapply+0x334/0x460 >>>> [xfs] >>>> [10919.516774] [<ffffffffb991432d>] wait_for_completion+0xfd/0x140 >>>> [10919.516782] [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 >>>> [10919.516821] [<ffffffffc05eb0a3>] ? _xfs_buf_read+0x23/0x40 [xfs] >>>> [10919.516859] [<ffffffffc05eafa9>] xfs_buf_submit_wait+0xf9/0x1d0 >>>> [xfs] >>>> [10919.516902] [<ffffffffc061b279>] ? xfs_trans_read_buf_map+0x199/0x400 >>>> [xfs] >>>> [10919.516940] [<ffffffffc05eb0a3>] _xfs_buf_read+0x23/0x40 [xfs] >>>> [10919.516977] [<ffffffffc05eb1b9>] xfs_buf_read_map+0xf9/0x160 [xfs] >>>> [10919.517022] [<ffffffffc061b279>] xfs_trans_read_buf_map+0x199/0x400 >>>> [xfs] >>>> [10919.517057] [<ffffffffc05c8d04>] xfs_da_read_buf+0xd4/0x100 [xfs] >>>> [10919.517091] [<ffffffffc05c8d53>] xfs_da3_node_read+0x23/0xd0 [xfs] >>>> [10919.517126] [<ffffffffc05c9fee>] xfs_da3_node_lookup_int+0x6e/0x2f0 >>>> [xfs] >>>> [10919.517160] [<ffffffffc05d5a1d>] xfs_dir2_node_lookup+0x4d/0x170 >>>> [xfs] >>>> [10919.517194] [<ffffffffc05ccf5d>] xfs_dir_lookup+0x1bd/0x1e0 [xfs] >>>> [10919.517233] [<ffffffffc05fd8d9>] xfs_lookup+0x69/0x140 [xfs] >>>> [10919.517271] [<ffffffffc05fa018>] xfs_vn_lookup+0x78/0xc0 [xfs] >>>> [10919.517278] [<ffffffffb9425cf3>] lookup_real+0x23/0x60 >>>> [10919.517283] [<ffffffffb9426702>] __lookup_hash+0x42/0x60 >>>> [10919.517288] [<ffffffffb942d519>] SYSC_renameat2+0x3a9/0x5a0 >>>> [10919.517296] [<ffffffffb94d3753>] ? selinux_file_free_security+0x2 >>>> 3/0x30 >>>> [10919.517304] [<ffffffffb992077b>] ? system_call_after_swapgs+0xc8/ >>>> 0x160 >>>> [10919.517309] [<ffffffffb992076f>] ? system_call_after_swapgs+0xbc/ >>>> 0x160 >>>> [10919.517313] [<ffffffffb992077b>] ? system_call_after_swapgs+0xc8/ >>>> 0x160 >>>> [10919.517318] [<ffffffffb992076f>] ? system_call_after_swapgs+0xbc/ >>>> 0x160 >>>> [10919.517323] [<ffffffffb942e58e>] SyS_renameat2+0xe/0x10 >>>> [10919.517328] [<ffffffffb942e5ce>] SyS_rename+0x1e/0x20 >>>> [10919.517333] [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 >>>> [10919.517339] [<ffffffffb992077b>] ? system_call_after_swapgs+0xc8/ >>>> 0x160 >>>> [11159.496095] INFO: task glusteriotwr9:15482 blocked for more than 120 >>>> seconds. >>>> [11159.497546] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" >>>> disables this message. >>>> [11159.498978] glusteriotwr9 D ffff971fa0fa1fa0 0 15482 1 >>>> 0x00000080 >>>> [11159.498984] Call Trace: >>>> [11159.498995] [<ffffffffb9911f00>] ? bit_wait+0x50/0x50 >>>> [11159.498999] [<ffffffffb9913f79>] schedule+0x29/0x70 >>>> [11159.499003] [<ffffffffb99118e9>] schedule_timeout+0x239/0x2c0 >>>> [11159.499056] [<ffffffffc05dd9b7>] ? xfs_iext_bno_to_ext+0xa7/0x1a0 >>>> [xfs] >>>> [11159.499082] [<ffffffffc05dd43e>] ? xfs_iext_bno_to_irec+0x8e/0xd0 >>>> [xfs] >>>> [11159.499090] [<ffffffffb92f7a12>] ? ktime_get_ts64+0x52/0xf0 >>>> [11159.499093] [<ffffffffb9911f00>] ? bit_wait+0x50/0x50 >>>> [11159.499097] [<ffffffffb991348d>] io_schedule_timeout+0xad/0x130 >>>> [11159.499101] [<ffffffffb9913528>] io_schedule+0x18/0x20 >>>> [11159.499104] [<ffffffffb9911f11>] bit_wait_io+0x11/0x50 >>>> [11159.499107] [<ffffffffb9911ac1>] __wait_on_bit_lock+0x61/0xc0 >>>> [11159.499113] [<ffffffffb9393634>] __lock_page+0x74/0x90 >>>> [11159.499118] [<ffffffffb92bc210>] ? wake_bit_function+0x40/0x40 >>>> [11159.499121] [<ffffffffb9394154>] __find_lock_page+0x54/0x70 >>>> [11159.499125] [<ffffffffb9394e85>] grab_cache_page_write_begin+0x >>>> 55/0xc0 >>>> [11159.499130] [<ffffffffb9484b76>] iomap_write_begin+0x66/0x100 >>>> [11159.499135] [<ffffffffb9484edf>] iomap_write_actor+0xcf/0x1d0 >>>> [11159.499140] [<ffffffffb9484e10>] ? iomap_write_end+0x80/0x80 >>>> [11159.499144] [<ffffffffb94854e7>] iomap_apply+0xb7/0x150 >>>> [11159.499149] [<ffffffffb9485621>] iomap_file_buffered_write+0xa1 >>>> /0xe0 >>>> [11159.499153] [<ffffffffb9484e10>] ? iomap_write_end+0x80/0x80 >>>> [11159.499182] [<ffffffffc05f025d>] >>>> xfs_file_buffered_aio_write+0x12d/0x2c0 >>>> [xfs] >>>> [11159.499213] [<ffffffffc05f057d>] xfs_file_aio_write+0x18d/0x1b0 >>>> [xfs] >>>> [11159.499217] [<ffffffffb941a533>] do_sync_write+0x93/0xe0 >>>> [11159.499222] [<ffffffffb941b010>] vfs_write+0xc0/0x1f0 >>>> [11159.499225] [<ffffffffb941c002>] SyS_pwrite64+0x92/0xc0 >>>> [11159.499230] [<ffffffffb992076f>] ? system_call_after_swapgs+0xbc/ >>>> 0x160 >>>> [11159.499234] [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 >>>> [11159.499238] [<ffffffffb992077b>] ? system_call_after_swapgs+0xc8/ >>>> 0x160 >>>> [11279.488720] INFO: task xfsaild/dm-10:1134 blocked for more than 120 >>>> seconds. >>>> [11279.490197] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" >>>> disables this message. >>>> [11279.491665] xfsaild/dm-10 D ffff9720a8660fd0 0 1134 2 >>>> 0x00000000 >>>> [11279.491671] Call Trace: >>>> [11279.491682] [<ffffffffb92a3a2e>] ? try_to_del_timer_sync+0x5e/0x90 >>>> [11279.491688] [<ffffffffb9913f79>] schedule+0x29/0x70 >>>> [11279.491744] [<ffffffffc060de36>] _xfs_log_force+0x1c6/0x2c0 [xfs] >>>> [11279.491750] [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 >>>> [11279.491783] [<ffffffffc0619fec>] ? xfsaild+0x16c/0x6f0 [xfs] >>>> [11279.491817] [<ffffffffc060df5c>] xfs_log_force+0x2c/0x70 [xfs] >>>> [11279.491849] [<ffffffffc0619e80>] ? xfs_trans_ail_cursor_first+0x90/0x90 >>>> [xfs] >>>> [11279.491880] [<ffffffffc0619fec>] xfsaild+0x16c/0x6f0 [xfs] >>>> [11279.491913] [<ffffffffc0619e80>] ? xfs_trans_ail_cursor_first+0x90/0x90 >>>> [xfs] >>>> [11279.491919] [<ffffffffb92bb161>] kthread+0xd1/0xe0 >>>> [11279.491926] [<ffffffffb92bb090>] ? insert_kthread_work+0x40/0x40 >>>> [11279.491932] [<ffffffffb9920677>] ret_from_fork_nospec_begin+0x2 >>>> 1/0x21 >>>> [11279.491936] [<ffffffffb92bb090>] ? insert_kthread_work+0x40/0x40 >>>> [11279.491976] INFO: task glusterclogfsyn:14934 blocked for more than >>>> 120 seconds. >>>> [11279.493466] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" >>>> disables this message. >>>> [11279.494952] glusterclogfsyn D ffff97209832af70 0 14934 1 >>>> 0x00000080 >>>> [11279.494957] Call Trace: >>>> [11279.494979] [<ffffffffc0309839>] ? __split_and_process_bio+0x2e9/0x520 >>>> [dm_mod] >>>> [11279.494983] [<ffffffffb9913f79>] schedule+0x29/0x70 >>>> [11279.494987] [<ffffffffb99118e9>] schedule_timeout+0x239/0x2c0 >>>> [11279.494997] [<ffffffffc0309d98>] ? dm_make_request+0x128/0x1a0 >>>> [dm_mod] >>>> [11279.495001] [<ffffffffb991348d>] io_schedule_timeout+0xad/0x130 >>>> [11279.495005] [<ffffffffb99145ad>] wait_for_completion_io+0xfd/0x140 >>>> [11279.495010] [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 >>>> [11279.495016] [<ffffffffb951e574>] blkdev_issue_flush+0xb4/0x110 >>>> [11279.495049] [<ffffffffc06064b9>] xfs_blkdev_issue_flush+0x19/0x20 >>>> [xfs] >>>> [11279.495079] [<ffffffffc05eec40>] xfs_file_fsync+0x1b0/0x1e0 [xfs] >>>> [11279.495086] [<ffffffffb944f0e7>] do_fsync+0x67/0xb0 >>>> [11279.495090] [<ffffffffb992076f>] ? system_call_after_swapgs+0xbc/ >>>> 0x160 >>>> [11279.495094] [<ffffffffb944f3d0>] SyS_fsync+0x10/0x20 >>>> [11279.495098] [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 >>>> [11279.495102] [<ffffffffb992077b>] ? system_call_after_swapgs+0xc8/ >>>> 0x160 >>>> [11279.495105] INFO: task glusterposixfsy:14941 blocked for more than >>>> 120 seconds. >>>> [11279.496606] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" >>>> disables this message. >>>> [11279.498114] glusterposixfsy D ffff972495f84f10 0 14941 1 >>>> 0x00000080 >>>> [11279.498118] Call Trace: >>>> [11279.498134] [<ffffffffc0309839>] ? __split_and_process_bio+0x2e9/0x520 >>>> [dm_mod] >>>> [11279.498138] [<ffffffffb9913f79>] schedule+0x29/0x70 >>>> [11279.498142] [<ffffffffb99118e9>] schedule_timeout+0x239/0x2c0 >>>> [11279.498152] [<ffffffffc0309d98>] ? dm_make_request+0x128/0x1a0 >>>> [dm_mod] >>>> [11279.498156] [<ffffffffb991348d>] io_schedule_timeout+0xad/0x130 >>>> [11279.498160] [<ffffffffb99145ad>] wait_for_completion_io+0xfd/0x140 >>>> [11279.498165] [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 >>>> [11279.498169] [<ffffffffb951e574>] blkdev_issue_flush+0xb4/0x110 >>>> [11279.498202] [<ffffffffc06064b9>] xfs_blkdev_issue_flush+0x19/0x20 >>>> [xfs] >>>> [11279.498231] [<ffffffffc05eec40>] xfs_file_fsync+0x1b0/0x1e0 [xfs] >>>> [11279.498238] [<ffffffffb944f0e7>] do_fsync+0x67/0xb0 >>>> [11279.498242] [<ffffffffb992076f>] ? system_call_after_swapgs+0xbc/ >>>> 0x160 >>>> [11279.498246] [<ffffffffb944f3f3>] SyS_fdatasync+0x13/0x20 >>>> [11279.498250] [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 >>>> [11279.498254] [<ffffffffb992077b>] ? system_call_after_swapgs+0xc8/ >>>> 0x160 >>>> [11279.498257] INFO: task glusteriotwr1:14950 blocked for more than 120 >>>> seconds. >>>> [11279.499789] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" >>>> disables this message. >>>> [11279.501343] glusteriotwr1 D ffff97208b6daf70 0 14950 1 >>>> 0x00000080 >>>> [11279.501348] Call Trace: >>>> [11279.501353] [<ffffffffb9913f79>] schedule+0x29/0x70 >>>> [11279.501390] [<ffffffffc060e388>] _xfs_log_force_lsn+0x2e8/0x340 >>>> [xfs] >>>> [11279.501396] [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 >>>> [11279.501428] [<ffffffffc05eeb97>] xfs_file_fsync+0x107/0x1e0 [xfs] >>>> [11279.501432] [<ffffffffb944ef3f>] generic_write_sync+0x4f/0x70 >>>> [11279.501461] [<ffffffffc05f0545>] xfs_file_aio_write+0x155/0x1b0 >>>> [xfs] >>>> [11279.501466] [<ffffffffb941a533>] do_sync_write+0x93/0xe0 >>>> [11279.501471] [<ffffffffb941b010>] vfs_write+0xc0/0x1f0 >>>> [11279.501475] [<ffffffffb941c002>] SyS_pwrite64+0x92/0xc0 >>>> [11279.501479] [<ffffffffb992076f>] ? system_call_after_swapgs+0xbc/ >>>> 0x160 >>>> [11279.501483] [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 >>>> [11279.501489] [<ffffffffb992077b>] ? system_call_after_swapgs+0xc8/ >>>> 0x160 >>>> [11279.501493] INFO: task glusteriotwr4:14953 blocked for more than 120 >>>> seconds. >>>> [11279.503047] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" >>>> disables this message. >>>> [11279.504630] glusteriotwr4 D ffff972499f2bf40 0 14953 1 >>>> 0x00000080 >>>> [11279.504635] Call Trace: >>>> [11279.504640] [<ffffffffb9913f79>] schedule+0x29/0x70 >>>> [11279.504676] [<ffffffffc060e388>] _xfs_log_force_lsn+0x2e8/0x340 >>>> [xfs] >>>> [11279.504681] [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 >>>> [11279.504710] [<ffffffffc05eeb97>] xfs_file_fsync+0x107/0x1e0 [xfs] >>>> [11279.504714] [<ffffffffb944f0e7>] do_fsync+0x67/0xb0 >>>> [11279.504718] [<ffffffffb992076f>] ? system_call_after_swapgs+0xbc/ >>>> 0x160 >>>> [11279.504722] [<ffffffffb944f3d0>] SyS_fsync+0x10/0x20 >>>> [11279.504725] [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 >>>> [11279.504730] [<ffffffffb992077b>] ? system_call_after_swapgs+0xc8/ >>>> 0x160 >>>> [12127.466494] perf: interrupt took too long (8263 > 8150), lowering >>>> kernel.perf_event_max_sample_rate to 24000 >>>> >>>> -------------------- >>>> I think this is the cause of the massive ovirt performance issues >>>> irrespective of gluster volume. At the time this happened, I was also >>>> ssh'ed into the host, and was doing some rpm querry commands. I had just >>>> run rpm -qa |grep glusterfs (to verify what version was actually >>>> installed), and that command took almost 2 minutes to return! Normally it >>>> takes less than 2 seconds. That is all pure local SSD IO, too.... >>>> >>>> I'm no expert, but its my understanding that anytime a software causes >>>> these kinds of issues, its a serious bug in the software, even if its >>>> mis-handled exceptions. Is this correct? >>>> >>>> --Jim >>>> >>>> On Tue, May 29, 2018 at 3:01 PM, Jim Kusznir <[email protected]> >>>> wrote: >>>> >>>>> I think this is the profile information for one of the volumes that >>>>> lives on the SSDs and is fully operational with no down/problem disks: >>>>> >>>>> [root@ovirt2 yum.repos.d]# gluster volume profile data info >>>>> Brick: ovirt2.nwfiber.com:/gluster/brick2/data >>>>> ---------------------------------------------- >>>>> Cumulative Stats: >>>>> Block Size: 256b+ 512b+ >>>>> 1024b+ >>>>> No. of Reads: 983 2696 >>>>> 1059 >>>>> No. of Writes: 0 1113 >>>>> 302 >>>>> >>>>> Block Size: 2048b+ 4096b+ >>>>> 8192b+ >>>>> No. of Reads: 852 88608 >>>>> 53526 >>>>> No. of Writes: 522 812340 >>>>> 76257 >>>>> >>>>> Block Size: 16384b+ 32768b+ >>>>> 65536b+ >>>>> No. of Reads: 54351 241901 >>>>> 15024 >>>>> No. of Writes: 21636 8656 >>>>> 8976 >>>>> >>>>> Block Size: 131072b+ >>>>> No. of Reads: 524156 >>>>> No. of Writes: 296071 >>>>> %-latency Avg-latency Min-Latency Max-Latency No. of calls >>>>> Fop >>>>> --------- ----------- ----------- ----------- ------------ >>>>> ---- >>>>> 0.00 0.00 us 0.00 us 0.00 us 4189 >>>>> RELEASE >>>>> 0.00 0.00 us 0.00 us 0.00 us 1257 >>>>> RELEASEDIR >>>>> 0.00 46.19 us 12.00 us 187.00 us 69 >>>>> FLUSH >>>>> 0.00 147.00 us 78.00 us 367.00 us 86 >>>>> REMOVEXATTR >>>>> 0.00 223.46 us 24.00 us 1166.00 us 149 >>>>> READDIR >>>>> 0.00 565.34 us 76.00 us 3639.00 us 88 >>>>> FTRUNCATE >>>>> 0.00 263.28 us 20.00 us 28385.00 us 228 >>>>> LK >>>>> 0.00 98.84 us 2.00 us 880.00 us 1198 >>>>> OPENDIR >>>>> 0.00 91.59 us 26.00 us 10371.00 us 3853 >>>>> STATFS >>>>> 0.00 494.14 us 17.00 us 193439.00 us 1171 >>>>> GETXATTR >>>>> 0.00 299.42 us 35.00 us 9799.00 us 2044 >>>>> READDIRP >>>>> 0.00 1965.31 us 110.00 us 382258.00 us 321 >>>>> XATTROP >>>>> 0.01 113.40 us 24.00 us 61061.00 us 8134 >>>>> STAT >>>>> 0.01 755.38 us 57.00 us 607603.00 us 3196 >>>>> DISCARD >>>>> 0.05 2690.09 us 58.00 us 2704761.00 us 3206 >>>>> OPEN >>>>> 0.10 119978.25 us 97.00 us 9406684.00 us 154 >>>>> SETATTR >>>>> 0.18 101.73 us 28.00 us 700477.00 us 313379 >>>>> FSTAT >>>>> 0.23 1059.84 us 25.00 us 2716124.00 us 38255 >>>>> LOOKUP >>>>> 0.47 1024.11 us 54.00 us 6197164.00 us 81455 >>>>> FXATTROP >>>>> 1.72 2984.00 us 15.00 us 37098954.00 us 103020 >>>>> FINODELK >>>>> 5.92 44315.32 us 51.00 us 24731536.00 us 23957 >>>>> FSYNC >>>>> 13.27 2399.78 us 25.00 us 22089540.00 us 991005 >>>>> READ >>>>> 37.00 5980.43 us 52.00 us 22099889.00 us 1108976 >>>>> WRITE >>>>> 41.04 5452.75 us 13.00 us 22102452.00 us 1349053 >>>>> INODELK >>>>> >>>>> Duration: 10026 seconds >>>>> Data Read: 80046027759 bytes >>>>> Data Written: 44496632320 bytes >>>>> >>>>> Interval 1 Stats: >>>>> Block Size: 256b+ 512b+ >>>>> 1024b+ >>>>> No. of Reads: 983 2696 >>>>> 1059 >>>>> No. of Writes: 0 838 >>>>> 185 >>>>> >>>>> Block Size: 2048b+ 4096b+ >>>>> 8192b+ >>>>> No. of Reads: 852 85856 >>>>> 51575 >>>>> No. of Writes: 382 705802 >>>>> 57812 >>>>> >>>>> Block Size: 16384b+ 32768b+ >>>>> 65536b+ >>>>> No. of Reads: 52673 232093 >>>>> 14984 >>>>> No. of Writes: 13499 4908 >>>>> 4242 >>>>> >>>>> Block Size: 131072b+ >>>>> No. of Reads: 460040 >>>>> No. of Writes: 6411 >>>>> %-latency Avg-latency Min-Latency Max-Latency No. of calls >>>>> Fop >>>>> --------- ----------- ----------- ----------- ------------ >>>>> ---- >>>>> 0.00 0.00 us 0.00 us 0.00 us 2093 >>>>> RELEASE >>>>> 0.00 0.00 us 0.00 us 0.00 us 1093 >>>>> RELEASEDIR >>>>> 0.00 53.38 us 26.00 us 111.00 us 16 >>>>> FLUSH >>>>> 0.00 145.14 us 78.00 us 367.00 us 71 >>>>> REMOVEXATTR >>>>> 0.00 190.96 us 114.00 us 298.00 us 71 >>>>> SETATTR >>>>> 0.00 213.38 us 24.00 us 1145.00 us 90 >>>>> READDIR >>>>> 0.00 263.28 us 20.00 us 28385.00 us 228 >>>>> LK >>>>> 0.00 101.76 us 2.00 us 880.00 us 1093 >>>>> OPENDIR >>>>> 0.01 93.60 us 27.00 us 10371.00 us 3090 >>>>> STATFS >>>>> 0.02 537.47 us 17.00 us 193439.00 us 1038 >>>>> GETXATTR >>>>> 0.03 297.44 us 35.00 us 9799.00 us 1990 >>>>> READDIRP >>>>> 0.03 2357.28 us 110.00 us 382258.00 us 253 >>>>> XATTROP >>>>> 0.04 385.93 us 58.00 us 47593.00 us 2091 >>>>> OPEN >>>>> 0.04 114.86 us 24.00 us 61061.00 us 7715 >>>>> STAT >>>>> 0.06 444.59 us 57.00 us 333240.00 us 3053 >>>>> DISCARD >>>>> 0.42 316.24 us 25.00 us 290728.00 us 29823 >>>>> LOOKUP >>>>> 0.73 257.92 us 54.00 us 344812.00 us 63296 >>>>> FXATTROP >>>>> 1.37 98.30 us 28.00 us 67621.00 us 313172 >>>>> FSTAT >>>>> 1.58 2124.69 us 51.00 us 849200.00 us 16717 >>>>> FSYNC >>>>> 5.73 162.46 us 52.00 us 748492.00 us 794079 >>>>> WRITE >>>>> 7.19 2065.17 us 16.00 us 37098954.00 us 78381 >>>>> FINODELK >>>>> 36.44 886.32 us 25.00 us 2216436.00 us 925421 >>>>> READ >>>>> 46.30 1178.04 us 13.00 us 1700704.00 us 884635 >>>>> INODELK >>>>> >>>>> Duration: 7485 seconds >>>>> Data Read: 71250527215 bytes >>>>> Data Written: 5119903744 bytes >>>>> >>>>> Brick: ovirt3.nwfiber.com:/gluster/brick2/data >>>>> ---------------------------------------------- >>>>> Cumulative Stats: >>>>> Block Size: 1b+ >>>>> No. of Reads: 0 >>>>> No. of Writes: 3264419 >>>>> %-latency Avg-latency Min-Latency Max-Latency No. of calls >>>>> Fop >>>>> --------- ----------- ----------- ----------- ------------ >>>>> ---- >>>>> 0.00 0.00 us 0.00 us 0.00 us 90 >>>>> FORGET >>>>> 0.00 0.00 us 0.00 us 0.00 us 9462 >>>>> RELEASE >>>>> 0.00 0.00 us 0.00 us 0.00 us 4254 >>>>> RELEASEDIR >>>>> 0.00 50.52 us 13.00 us 190.00 us 71 >>>>> FLUSH >>>>> 0.00 186.97 us 87.00 us 713.00 us 86 >>>>> REMOVEXATTR >>>>> 0.00 79.32 us 33.00 us 189.00 us 228 >>>>> LK >>>>> 0.00 220.98 us 129.00 us 513.00 us 86 >>>>> SETATTR >>>>> 0.01 259.30 us 26.00 us 2632.00 us 137 >>>>> READDIR >>>>> 0.02 322.76 us 145.00 us 2125.00 us 321 >>>>> XATTROP >>>>> 0.03 109.55 us 2.00 us 1258.00 us 1193 >>>>> OPENDIR >>>>> 0.05 70.21 us 21.00 us 431.00 us 3196 >>>>> DISCARD >>>>> 0.05 169.26 us 21.00 us 2315.00 us 1545 >>>>> GETXATTR >>>>> 0.12 176.85 us 63.00 us 2844.00 us 3206 >>>>> OPEN >>>>> 0.61 303.49 us 90.00 us 3085.00 us 9633 >>>>> FSTAT >>>>> 2.44 305.66 us 28.00 us 3716.00 us 38230 >>>>> LOOKUP >>>>> 4.52 266.22 us 55.00 us 53424.00 us 81455 >>>>> FXATTROP >>>>> 6.96 1397.99 us 51.00 us 64822.00 us 23889 >>>>> FSYNC >>>>> 16.48 84.74 us 25.00 us 6917.00 us 932592 >>>>> WRITE >>>>> 30.16 106.90 us 13.00 us 3920189.00 us 1353046 >>>>> INODELK >>>>> 38.55 1794.52 us 14.00 us 16210553.00 us 103039 >>>>> FINODELK >>>>> >>>>> Duration: 66562 seconds >>>>> Data Read: 0 bytes >>>>> Data Written: 3264419 bytes >>>>> >>>>> Interval 1 Stats: >>>>> Block Size: 1b+ >>>>> No. of Reads: 0 >>>>> No. of Writes: 794080 >>>>> %-latency Avg-latency Min-Latency Max-Latency No. of calls >>>>> Fop >>>>> --------- ----------- ----------- ----------- ------------ >>>>> ---- >>>>> 0.00 0.00 us 0.00 us 0.00 us 2093 >>>>> RELEASE >>>>> 0.00 0.00 us 0.00 us 0.00 us 1093 >>>>> RELEASEDIR >>>>> 0.00 70.31 us 26.00 us 125.00 us 16 >>>>> FLUSH >>>>> 0.00 193.10 us 103.00 us 713.00 us 71 >>>>> REMOVEXATTR >>>>> 0.01 227.32 us 133.00 us 513.00 us 71 >>>>> SETATTR >>>>> 0.01 79.32 us 33.00 us 189.00 us 228 >>>>> LK >>>>> 0.01 259.83 us 35.00 us 1138.00 us 89 >>>>> READDIR >>>>> 0.03 318.26 us 145.00 us 2047.00 us 253 >>>>> XATTROP >>>>> 0.04 112.67 us 3.00 us 1258.00 us 1093 >>>>> OPENDIR >>>>> 0.06 167.98 us 23.00 us 1951.00 us 1014 >>>>> GETXATTR >>>>> 0.08 70.97 us 22.00 us 431.00 us 3053 >>>>> DISCARD >>>>> 0.13 183.78 us 66.00 us 2844.00 us 2091 >>>>> OPEN >>>>> 1.01 303.82 us 90.00 us 3085.00 us 9610 >>>>> FSTAT >>>>> 3.27 316.59 us 30.00 us 3716.00 us 29820 >>>>> LOOKUP >>>>> 5.83 265.79 us 59.00 us 53424.00 us 63296 >>>>> FXATTROP >>>>> 7.95 1373.89 us 51.00 us 64822.00 us 16717 >>>>> FSYNC >>>>> 23.17 851.99 us 14.00 us 16210553.00 us 78555 >>>>> FINODELK >>>>> 24.04 87.44 us 27.00 us 6917.00 us 794081 >>>>> WRITE >>>>> 34.36 111.91 us 14.00 us 984871.00 us 886790 >>>>> INODELK >>>>> >>>>> Duration: 7485 seconds >>>>> Data Read: 0 bytes >>>>> Data Written: 794080 bytes >>>>> >>>>> >>>>> ----------------------- >>>>> Here is the data from the volume that is backed by the SHDDs and has >>>>> one failed disk: >>>>> [root@ovirt2 yum.repos.d]# gluster volume profile data-hdd info >>>>> Brick: 172.172.1.12:/gluster/brick3/data-hdd >>>>> -------------------------------------------- >>>>> Cumulative Stats: >>>>> Block Size: 256b+ 512b+ >>>>> 1024b+ >>>>> No. of Reads: 1702 86 >>>>> 16 >>>>> No. of Writes: 0 767 >>>>> 71 >>>>> >>>>> Block Size: 2048b+ 4096b+ >>>>> 8192b+ >>>>> No. of Reads: 19 51841 >>>>> 2049 >>>>> No. of Writes: 76 60668 >>>>> 35727 >>>>> >>>>> Block Size: 16384b+ 32768b+ >>>>> 65536b+ >>>>> No. of Reads: 1744 639 >>>>> 1088 >>>>> No. of Writes: 8524 2410 >>>>> 1285 >>>>> >>>>> Block Size: 131072b+ >>>>> No. of Reads: 771999 >>>>> No. of Writes: 29584 >>>>> %-latency Avg-latency Min-Latency Max-Latency No. of calls >>>>> Fop >>>>> --------- ----------- ----------- ----------- ------------ >>>>> ---- >>>>> 0.00 0.00 us 0.00 us 0.00 us 2902 >>>>> RELEASE >>>>> 0.00 0.00 us 0.00 us 0.00 us 1517 >>>>> RELEASEDIR >>>>> 0.00 197.00 us 197.00 us 197.00 us 1 >>>>> FTRUNCATE >>>>> 0.00 70.24 us 16.00 us 758.00 us 51 >>>>> FLUSH >>>>> 0.00 143.93 us 82.00 us 305.00 us 57 >>>>> REMOVEXATTR >>>>> 0.00 178.63 us 105.00 us 712.00 us 60 >>>>> SETATTR >>>>> 0.00 67.30 us 19.00 us 572.00 us 555 >>>>> LK >>>>> 0.00 322.80 us 23.00 us 4673.00 us 138 >>>>> READDIR >>>>> 0.00 336.56 us 106.00 us 11994.00 us 237 >>>>> XATTROP >>>>> 0.00 84.70 us 28.00 us 1071.00 us 3469 >>>>> STATFS >>>>> 0.01 387.75 us 2.00 us 146017.00 us 1467 >>>>> OPENDIR >>>>> 0.01 148.59 us 21.00 us 64374.00 us 4454 >>>>> STAT >>>>> 0.02 783.02 us 16.00 us 93502.00 us 1902 >>>>> GETXATTR >>>>> 0.03 1516.10 us 17.00 us 210690.00 us 1364 >>>>> ENTRYLK >>>>> 0.03 2555.47 us 300.00 us 674454.00 us 1064 >>>>> READDIRP >>>>> 0.07 85.74 us 19.00 us 68340.00 us 62849 >>>>> FSTAT >>>>> 0.07 1978.12 us 59.00 us 202596.00 us 2729 >>>>> OPEN >>>>> 0.22 708.57 us 15.00 us 394799.00 us 25447 >>>>> LOOKUP >>>>> 5.94 2331.74 us 15.00 us 1099530.00 us 207534 >>>>> FINODELK >>>>> 7.31 8311.75 us 58.00 us 1800216.00 us 71668 >>>>> FXATTROP >>>>> 12.49 7735.19 us 51.00 us 3595513.00 us 131642 >>>>> WRITE >>>>> 17.70 957.08 us 16.00 us 13700466.00 us 1508160 >>>>> INODELK >>>>> 24.55 2546.43 us 26.00 us 5077347.00 us 786060 >>>>> READ >>>>> 31.56 49699.15 us 47.00 us 3746331.00 us 51777 >>>>> FSYNC >>>>> >>>>> Duration: 10101 seconds >>>>> Data Read: 101562897361 bytes >>>>> Data Written: 4834450432 bytes >>>>> >>>>> Interval 0 Stats: >>>>> Block Size: 256b+ 512b+ >>>>> 1024b+ >>>>> No. of Reads: 1702 86 >>>>> 16 >>>>> No. of Writes: 0 767 >>>>> 71 >>>>> >>>>> Block Size: 2048b+ 4096b+ >>>>> 8192b+ >>>>> No. of Reads: 19 51841 >>>>> 2049 >>>>> No. of Writes: 76 60668 >>>>> 35727 >>>>> >>>>> Block Size: 16384b+ 32768b+ >>>>> 65536b+ >>>>> No. of Reads: 1744 639 >>>>> 1088 >>>>> No. of Writes: 8524 2410 >>>>> 1285 >>>>> >>>>> Block Size: 131072b+ >>>>> No. of Reads: 771999 >>>>> No. of Writes: 29584 >>>>> %-latency Avg-latency Min-Latency Max-Latency No. of calls >>>>> Fop >>>>> --------- ----------- ----------- ----------- ------------ >>>>> ---- >>>>> 0.00 0.00 us 0.00 us 0.00 us 2902 >>>>> RELEASE >>>>> 0.00 0.00 us 0.00 us 0.00 us 1517 >>>>> RELEASEDIR >>>>> 0.00 197.00 us 197.00 us 197.00 us 1 >>>>> FTRUNCATE >>>>> 0.00 70.24 us 16.00 us 758.00 us 51 >>>>> FLUSH >>>>> 0.00 143.93 us 82.00 us 305.00 us 57 >>>>> REMOVEXATTR >>>>> 0.00 178.63 us 105.00 us 712.00 us 60 >>>>> SETATTR >>>>> 0.00 67.30 us 19.00 us 572.00 us 555 >>>>> LK >>>>> 0.00 322.80 us 23.00 us 4673.00 us 138 >>>>> READDIR >>>>> 0.00 336.56 us 106.00 us 11994.00 us 237 >>>>> XATTROP >>>>> 0.00 84.70 us 28.00 us 1071.00 us 3469 >>>>> STATFS >>>>> 0.01 387.75 us 2.00 us 146017.00 us 1467 >>>>> OPENDIR >>>>> 0.01 148.59 us 21.00 us 64374.00 us 4454 >>>>> STAT >>>>> 0.02 783.02 us 16.00 us 93502.00 us 1902 >>>>> GETXATTR >>>>> 0.03 1516.10 us 17.00 us 210690.00 us 1364 >>>>> ENTRYLK >>>>> 0.03 2555.47 us 300.00 us 674454.00 us 1064 >>>>> READDIRP >>>>> 0.07 85.73 us 19.00 us 68340.00 us 62849 >>>>> FSTAT >>>>> 0.07 1978.12 us 59.00 us 202596.00 us 2729 >>>>> OPEN >>>>> 0.22 708.57 us 15.00 us 394799.00 us 25447 >>>>> LOOKUP >>>>> 5.94 2334.57 us 15.00 us 1099530.00 us 207534 >>>>> FINODELK >>>>> 7.31 8311.49 us 58.00 us 1800216.00 us 71668 >>>>> FXATTROP >>>>> 12.49 7735.32 us 51.00 us 3595513.00 us 131642 >>>>> WRITE >>>>> 17.71 957.08 us 16.00 us 13700466.00 us 1508160 >>>>> INODELK >>>>> 24.56 2546.42 us 26.00 us 5077347.00 us 786060 >>>>> READ >>>>> 31.54 49651.63 us 47.00 us 3746331.00 us 51777 >>>>> FSYNC >>>>> >>>>> Duration: 10101 seconds >>>>> Data Read: 101562897361 bytes >>>>> Data Written: 4834450432 bytes >>>>> >>>>> >>>>> On Tue, May 29, 2018 at 2:55 PM, Jim Kusznir <[email protected]> >>>>> wrote: >>>>> >>>>>> Thank you for your response. >>>>>> >>>>>> I have 4 gluster volumes. 3 are replica 2 + arbitrator. replica >>>>>> bricks are on ovirt1 and ovirt2, arbitrator on ovirt3. The 4th volume is >>>>>> replica 3, with a brick on all three ovirt machines. >>>>>> >>>>>> The first 3 volumes are on an SSD disk; the 4th is on a Seagate SSHD >>>>>> (same in all three machines). On ovirt3, the SSHD has reported hard IO >>>>>> failures, and that brick is offline. However, the other two replicas are >>>>>> fully operational (although they still show contents in the heal info >>>>>> command that won't go away, but that may be the case until I replace the >>>>>> failed disk). >>>>>> >>>>>> What is bothering me is that ALL 4 gluster volumes are showing >>>>>> horrible performance issues. At this point, as the bad disk has been >>>>>> completely offlined, I would expect gluster to perform at normal speed, >>>>>> but >>>>>> that is definitely not the case. >>>>>> >>>>>> I've also noticed that the performance hits seem to come in waves: >>>>>> things seem to work acceptably (but slow) for a while, then suddenly, its >>>>>> as if all disk IO on all volumes (including non-gluster local OS disk >>>>>> volumes for the hosts) pause for about 30 seconds, then IO resumes again. >>>>>> During those times, I start getting VM not responding and host not >>>>>> responding notices as well as the applications having major issues. >>>>>> >>>>>> I've shut down most of my VMs and am down to just my essential core >>>>>> VMs (shedded about 75% of my VMs). I still am experiencing the same >>>>>> issues. >>>>>> >>>>>> Am I correct in believing that once the failed disk was brought >>>>>> offline that performance should return to normal? >>>>>> >>>>>> On Tue, May 29, 2018 at 1:27 PM, Alex K <[email protected]> >>>>>> wrote: >>>>>> >>>>>>> I would check disks status and accessibility of mount points where >>>>>>> your gluster volumes reside. >>>>>>> >>>>>>> On Tue, May 29, 2018, 22:28 Jim Kusznir <[email protected]> wrote: >>>>>>> >>>>>>>> On one ovirt server, I'm now seeing these messages: >>>>>>>> [56474.239725] blk_update_request: 63 callbacks suppressed >>>>>>>> [56474.239732] blk_update_request: I/O error, dev dm-2, sector 0 >>>>>>>> [56474.240602] blk_update_request: I/O error, dev dm-2, sector >>>>>>>> 3905945472 >>>>>>>> [56474.241346] blk_update_request: I/O error, dev dm-2, sector >>>>>>>> 3905945584 >>>>>>>> [56474.242236] blk_update_request: I/O error, dev dm-2, sector 2048 >>>>>>>> [56474.243072] blk_update_request: I/O error, dev dm-2, sector >>>>>>>> 3905943424 >>>>>>>> [56474.243997] blk_update_request: I/O error, dev dm-2, sector >>>>>>>> 3905943536 >>>>>>>> [56474.247347] blk_update_request: I/O error, dev dm-2, sector 0 >>>>>>>> [56474.248315] blk_update_request: I/O error, dev dm-2, sector >>>>>>>> 3905945472 >>>>>>>> [56474.249231] blk_update_request: I/O error, dev dm-2, sector >>>>>>>> 3905945584 >>>>>>>> [56474.250221] blk_update_request: I/O error, dev dm-2, sector 2048 >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Tue, May 29, 2018 at 11:59 AM, Jim Kusznir <[email protected]> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> I see in messages on ovirt3 (my 3rd machine, the one upgraded to >>>>>>>>> 4.2): >>>>>>>>> >>>>>>>>> May 29 11:54:41 ovirt3 ovs-vsctl: >>>>>>>>> ovs|00001|db_ctl_base|ERR|unix:/var/run/openvswitch/db.sock: >>>>>>>>> database connection failed (No such file or directory) >>>>>>>>> May 29 11:54:51 ovirt3 ovs-vsctl: >>>>>>>>> ovs|00001|db_ctl_base|ERR|unix:/var/run/openvswitch/db.sock: >>>>>>>>> database connection failed (No such file or directory) >>>>>>>>> May 29 11:55:01 ovirt3 ovs-vsctl: >>>>>>>>> ovs|00001|db_ctl_base|ERR|unix:/var/run/openvswitch/db.sock: >>>>>>>>> database connection failed (No such file or directory) >>>>>>>>> (appears a lot). >>>>>>>>> >>>>>>>>> I also found on the ssh session of that, some sysv warnings about >>>>>>>>> the backing disk for one of the gluster volumes (straight replica 3). >>>>>>>>> The >>>>>>>>> glusterfs process for that disk on that machine went offline. Its my >>>>>>>>> understanding that it should continue to work with the other two >>>>>>>>> machines >>>>>>>>> while I attempt to replace that disk, right? Attempted writes >>>>>>>>> (touching an >>>>>>>>> empty file) can take 15 seconds, repeating it later will be much >>>>>>>>> faster. >>>>>>>>> >>>>>>>>> Gluster generates a bunch of different log files, I don't know >>>>>>>>> what ones you want, or from which machine(s). >>>>>>>>> >>>>>>>>> How do I do "volume profiling"? >>>>>>>>> >>>>>>>>> Thanks! >>>>>>>>> >>>>>>>>> On Tue, May 29, 2018 at 11:53 AM, Sahina Bose <[email protected]> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Do you see errors reported in the mount logs for the volume? If >>>>>>>>>> so, could you attach the logs? >>>>>>>>>> Any issues with your underlying disks. Can you also attach output >>>>>>>>>> of volume profiling? >>>>>>>>>> >>>>>>>>>> On Wed, May 30, 2018 at 12:13 AM, Jim Kusznir < >>>>>>>>>> [email protected]> wrote: >>>>>>>>>> >>>>>>>>>>> Ok, things have gotten MUCH worse this morning. I'm getting >>>>>>>>>>> random errors from VMs, right now, about a third of my VMs have >>>>>>>>>>> been paused >>>>>>>>>>> due to storage issues, and most of the remaining VMs are not >>>>>>>>>>> performing >>>>>>>>>>> well. >>>>>>>>>>> >>>>>>>>>>> At this point, I am in full EMERGENCY mode, as my production >>>>>>>>>>> services are now impacted, and I'm getting calls coming in with >>>>>>>>>>> problems... >>>>>>>>>>> >>>>>>>>>>> I'd greatly appreciate help...VMs are running VERY slowly (when >>>>>>>>>>> they run), and they are steadily getting worse. I don't know why. >>>>>>>>>>> I was >>>>>>>>>>> seeing CPU peaks (to 100%) on several VMs, in perfect sync, for a >>>>>>>>>>> few >>>>>>>>>>> minutes at a time (while the VM became unresponsive and any VMs I >>>>>>>>>>> was >>>>>>>>>>> logged into that were linux were giving me the CPU stuck messages >>>>>>>>>>> in my >>>>>>>>>>> origional post). Is all this storage related? >>>>>>>>>>> >>>>>>>>>>> I also have two different gluster volumes for VM storage, and >>>>>>>>>>> only one had the issues, but now VMs in both are being affected at >>>>>>>>>>> the same >>>>>>>>>>> time and same way. >>>>>>>>>>> >>>>>>>>>>> --Jim >>>>>>>>>>> >>>>>>>>>>> On Mon, May 28, 2018 at 10:50 PM, Sahina Bose <[email protected] >>>>>>>>>>> > wrote: >>>>>>>>>>> >>>>>>>>>>>> [Adding gluster-users to look at the heal issue] >>>>>>>>>>>> >>>>>>>>>>>> On Tue, May 29, 2018 at 9:17 AM, Jim Kusznir < >>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Hello: >>>>>>>>>>>>> >>>>>>>>>>>>> I've been having some cluster and gluster performance issues >>>>>>>>>>>>> lately. I also found that my cluster was out of date, and was >>>>>>>>>>>>> trying to >>>>>>>>>>>>> apply updates (hoping to fix some of these), and discovered the >>>>>>>>>>>>> ovirt 4.1 >>>>>>>>>>>>> repos were taken completely offline. So, I was forced to begin >>>>>>>>>>>>> an upgrade >>>>>>>>>>>>> to 4.2. According to docs I found/read, I needed only add the >>>>>>>>>>>>> new repo, do >>>>>>>>>>>>> a yum update, reboot, and be good on my hosts (did the yum >>>>>>>>>>>>> update, the >>>>>>>>>>>>> engine-setup on my hosted engine). Things seemed to work >>>>>>>>>>>>> relatively well, >>>>>>>>>>>>> except for a gluster sync issue that showed up. >>>>>>>>>>>>> >>>>>>>>>>>>> My cluster is a 3 node hyperconverged cluster. I upgraded the >>>>>>>>>>>>> hosted engine first, then engine 3. When engine 3 came back up, >>>>>>>>>>>>> for some >>>>>>>>>>>>> reason one of my gluster volumes would not sync. Here's sample >>>>>>>>>>>>> output: >>>>>>>>>>>>> >>>>>>>>>>>>> [root@ovirt3 ~]# gluster volume heal data-hdd info >>>>>>>>>>>>> Brick 172.172.1.11:/gluster/brick3/data-hdd >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/48d7ecb8-7ac5-4 >>>>>>>>>>>>> 725-bca5-b3519681cf2f/0d6080b0-7018-4fa3-bb82-1dd9ef07d9b9 >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/647be733-f153-4 >>>>>>>>>>>>> cdc-85bd-ba72544c2631/b453a300-0602-4be1-8310-8bd5abe00971 >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/6da854d1-b6be-4 >>>>>>>>>>>>> 46b-9bf0-90a0dbbea830/3c93bd1f-b7fa-4aa2-b445-6904e31839ba >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/7f647567-d18c-4 >>>>>>>>>>>>> 4f1-a58e-9b8865833acb/f9364470-9770-4bb1-a6b9-a54861849625 >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/f3c8e7aa-6ef2-4 >>>>>>>>>>>>> 2a7-93d4-e0a4df6dd2fa/2eb0b1ad-2606-44ef-9cd3-ae59610a504b >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/b1ea3f62-0f05-4 >>>>>>>>>>>>> ded-8c82-9c91c90e0b61/d5d6bf5a-499f-431d-9013-5453db93ed32 >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/8c8b5147-e9d6-4 >>>>>>>>>>>>> 810-b45b-185e3ed65727/16f08231-93b0-489d-a2fd-687b6bf88eaa >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/12924435-b9c2-4 >>>>>>>>>>>>> aab-ba19-1c1bc31310ef/07b3db69-440e-491e-854c-bbfa18a7cff2 >>>>>>>>>>>>> Status: Connected >>>>>>>>>>>>> Number of entries: 8 >>>>>>>>>>>>> >>>>>>>>>>>>> Brick 172.172.1.12:/gluster/brick3/data-hdd >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/48d7ecb8-7ac5-4 >>>>>>>>>>>>> 725-bca5-b3519681cf2f/0d6080b0-7018-4fa3-bb82-1dd9ef07d9b9 >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/647be733-f153-4 >>>>>>>>>>>>> cdc-85bd-ba72544c2631/b453a300-0602-4be1-8310-8bd5abe00971 >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/b1ea3f62-0f05-4 >>>>>>>>>>>>> ded-8c82-9c91c90e0b61/d5d6bf5a-499f-431d-9013-5453db93ed32 >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/6da854d1-b6be-4 >>>>>>>>>>>>> 46b-9bf0-90a0dbbea830/3c93bd1f-b7fa-4aa2-b445-6904e31839ba >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/7f647567-d18c-4 >>>>>>>>>>>>> 4f1-a58e-9b8865833acb/f9364470-9770-4bb1-a6b9-a54861849625 >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/8c8b5147-e9d6-4 >>>>>>>>>>>>> 810-b45b-185e3ed65727/16f08231-93b0-489d-a2fd-687b6bf88eaa >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/12924435-b9c2-4 >>>>>>>>>>>>> aab-ba19-1c1bc31310ef/07b3db69-440e-491e-854c-bbfa18a7cff2 >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/f3c8e7aa-6ef2-4 >>>>>>>>>>>>> 2a7-93d4-e0a4df6dd2fa/2eb0b1ad-2606-44ef-9cd3-ae59610a504b >>>>>>>>>>>>> Status: Connected >>>>>>>>>>>>> Number of entries: 8 >>>>>>>>>>>>> >>>>>>>>>>>>> Brick 172.172.1.13:/gluster/brick3/data-hdd >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/b1ea3f62-0f05-4 >>>>>>>>>>>>> ded-8c82-9c91c90e0b61/d5d6bf5a-499f-431d-9013-5453db93ed32 >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/8c8b5147-e9d6-4 >>>>>>>>>>>>> 810-b45b-185e3ed65727/16f08231-93b0-489d-a2fd-687b6bf88eaa >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/12924435-b9c2-4 >>>>>>>>>>>>> aab-ba19-1c1bc31310ef/07b3db69-440e-491e-854c-bbfa18a7cff2 >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/f3c8e7aa-6ef2-4 >>>>>>>>>>>>> 2a7-93d4-e0a4df6dd2fa/2eb0b1ad-2606-44ef-9cd3-ae59610a504b >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/647be733-f153-4 >>>>>>>>>>>>> cdc-85bd-ba72544c2631/b453a300-0602-4be1-8310-8bd5abe00971 >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/48d7ecb8-7ac5-4 >>>>>>>>>>>>> 725-bca5-b3519681cf2f/0d6080b0-7018-4fa3-bb82-1dd9ef07d9b9 >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/6da854d1-b6be-4 >>>>>>>>>>>>> 46b-9bf0-90a0dbbea830/3c93bd1f-b7fa-4aa2-b445-6904e31839ba >>>>>>>>>>>>> /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/7f647567-d18c-4 >>>>>>>>>>>>> 4f1-a58e-9b8865833acb/f9364470-9770-4bb1-a6b9-a54861849625 >>>>>>>>>>>>> Status: Connected >>>>>>>>>>>>> Number of entries: 8 >>>>>>>>>>>>> >>>>>>>>>>>>> --------- >>>>>>>>>>>>> Its been in this state for a couple days now, and bandwidth >>>>>>>>>>>>> monitoring shows no appreciable data moving. I've tried >>>>>>>>>>>>> repeatedly >>>>>>>>>>>>> commanding a full heal from all three clusters in the node. Its >>>>>>>>>>>>> always the >>>>>>>>>>>>> same files that need healing. >>>>>>>>>>>>> >>>>>>>>>>>>> When running gluster volume heal data-hdd statistics, I see >>>>>>>>>>>>> sometimes different information, but always some number of "heal >>>>>>>>>>>>> failed" >>>>>>>>>>>>> entries. It shows 0 for split brain. >>>>>>>>>>>>> >>>>>>>>>>>>> I'm not quite sure what to do. I suspect it may be due to >>>>>>>>>>>>> nodes 1 and 2 still being on the older ovirt/gluster release, but >>>>>>>>>>>>> I'm >>>>>>>>>>>>> afraid to upgrade and reboot them until I have a good gluster >>>>>>>>>>>>> sync (don't >>>>>>>>>>>>> need to create a split brain issue). How do I proceed with this? >>>>>>>>>>>>> >>>>>>>>>>>>> Second issue: I've been experiencing VERY POOR performance on >>>>>>>>>>>>> most of my VMs. To the tune that logging into a windows 10 vm >>>>>>>>>>>>> via remote >>>>>>>>>>>>> desktop can take 5 minutes, launching quickbooks inside said vm >>>>>>>>>>>>> can easily >>>>>>>>>>>>> take 10 minutes. On some linux VMs, I get random messages like >>>>>>>>>>>>> this: >>>>>>>>>>>>> Message from syslogd@unifi at May 28 20:39:23 ... >>>>>>>>>>>>> kernel:[6171996.308904] NMI watchdog: BUG: soft lockup - >>>>>>>>>>>>> CPU#0 stuck for 22s! [mongod:14766] >>>>>>>>>>>>> >>>>>>>>>>>>> (the process and PID are often different) >>>>>>>>>>>>> >>>>>>>>>>>>> I'm not quite sure what to do about this either. My initial >>>>>>>>>>>>> thought was upgrad everything to current and see if its still >>>>>>>>>>>>> there, but I >>>>>>>>>>>>> cannot move forward with that until my gluster is healed... >>>>>>>>>>>>> >>>>>>>>>>>>> Thanks! >>>>>>>>>>>>> --Jim >>>>>>>>>>>>> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> Users mailing list -- [email protected] >>>>>>>>>>>>> To unsubscribe send an email to [email protected] >>>>>>>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>>>>>>>> oVirt Code of Conduct: https://www.ovirt.org/communit >>>>>>>>>>>>> y/about/community-guidelines/ >>>>>>>>>>>>> List Archives: https://lists.ovirt.org/archiv >>>>>>>>>>>>> es/list/[email protected]/message/3LEV6ZQ3JV2XLAL7NYBTXOYMYUOT >>>>>>>>>>>>> IRQF/ >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> Users mailing list -- [email protected] >>>>>>>> To unsubscribe send an email to [email protected] >>>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>>> oVirt Code of Conduct: https://www.ovirt.org/communit >>>>>>>> y/about/community-guidelines/ >>>>>>>> List Archives: https://lists.ovirt.org/archiv >>>>>>>> es/list/[email protected]/message/ACO7RFSLBSRBAIONIC2HQ6Z24ZDES5MF/ >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> _______________________________________________ >> Users mailing list -- [email protected] >> To unsubscribe send an email to [email protected] >> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >> oVirt Code of Conduct: https://www.ovirt.org/community/about/community- >> guidelines/ >> List Archives: https://lists.ovirt.org/archives/list/[email protected]/ >> message/JKBSLSZFRMFGILXGZB4YAMJIJKEKNGMA/ >> > >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/OSKM4MFRYYLRRN2QOIF4FVXPX2UK3JU7/

