Re: [Gluster-users] Reliability issues with Gluster 3.10 and shard
On 15 May 2017 at 11:01, Benjamin Kingstonwrote: > I resolved this with the following settings, particularly disabling > features.ctr-enabled > That's odd. CTR should be enabled for tiered volumes. Was it enabled by default? > > olume Name: storage2 > Type: Distributed-Replicate > Volume ID: adaabca5-25ed-4e7f-ae86-2f20fc0143a8 > Status: Started > Snapshot Count: 0 > Number of Bricks: 3 x (2 + 1) = 9 > Transport-type: tcp > Bricks: > Brick1: fd00:0:0:3::6:/mnt/gluster/storage/brick0/glusterfs2 > Brick2: fd00:0:0:3::8:/mnt/gluster/storage/brick0/glusterfs2 > Brick3: fd00:0:0:3::10:/mnt/gluster/storage/brick0/glusterfs (arbiter) > Brick4: fd00:0:0:3::6:/mnt/gluster/storage/brick1/glusterfs2 > Brick5: fd00:0:0:3::8:/mnt/gluster/storage/brick1/glusterfs2 > Brick6: fd00:0:0:3::10:/mnt/gluster/storage/brick1/glusterfs (arbiter) > Brick7: fd00:0:0:3::6:/mnt/gluster/storage/brick2/glusterfs2 > Brick8: fd00:0:0:3::8:/mnt/gluster/storage/brick2/glusterfs2 > Brick9: fd00:0:0:3::10:/mnt/gluster/storage/brick2/glusterfs (arbiter) > Options Reconfigured: > performance.write-behind-window-size: 4MB > performance.cache-invalidation: on > transport.keepalive: on > performance.write-behind: on > performance.read-ahead: on > performance.io-cache: on > performance.stat-prefetch: on > performance.open-behind: on > cluster.use-compound-fops: on > performance.cache-ima-xattrs: on > features.cache-invalidation: on > client.event-threads: 4 > cluster.data-self-heal-algorithm: full > performance.client-io-threads: on > server.event-threads: 4 > performance.quick-read: on > features.scrub: Active > features.bitrot: on > features.shard: on > transport.address-family: inet6 > nfs.disable: on > server.allow-insecure: on > user.cifs: off > cluster.quorum-type: auto > cluster.server-quorum-type: server > cluster.tier-compact: on > diagnostics.brick-log-level: WARNING > diagnostics.client-log-level: WARNING > cluster.self-heal-daemon: enable > performance.cache-samba-metadata: on > cluster.brick-multiplex: off > cluster.enable-shared-storage: enable > nfs-ganesha: enable > > > > -ben > > On Sat, May 13, 2017 at 12:20 PM, Benjamin Kingston > wrote: > >> Hers's some log entries from nfs-ganesha gfapi >> >> [2017-05-13 19:02:54.105936] E [MSGID: 133010] >> [shard.c:1706:shard_common_lookup_shards_cbk] 0-storage2-shard: Lookup >> on shard 11 failed. Base file gfid = 1494c083-a618-4eba-80a0-147e656dd9d0 >> [Input/output error] >> [2017-05-13 19:02:54.106176] E [MSGID: 133010] >> [shard.c:1706:shard_common_lookup_shards_cbk] 0-storage2-shard: Lookup >> on shard 2 failed. Base file gfid = 1494c083-a618-4eba-80a0-147e656dd9d0 >> [Input/output error] >> [2017-05-13 19:02:54.106288] E [MSGID: 133010] >> [shard.c:1706:shard_common_lookup_shards_cbk] 0-storage2-shard: Lookup >> on shard 1 failed. Base file gfid = 1494c083-a618-4eba-80a0-147e656dd9d0 >> [Input/output error] >> [2017-05-13 19:02:54.384922] I [MSGID: 108026] >> [afr-self-heal-metadata.c:52:__afr_selfheal_metadata_do] >> 0-storage2-replicate-2: performing metadata selfheal on >> fe651475-226e-42a3-be2d-751d4f58e383 >> [2017-05-13 19:02:54.385894] W [MSGID: 114031] >> [client-rpc-fops.c:2258:client3_3_setattr_cbk] 0-storage2-client-8: >> remote operation failed [Operation not permitted] >> [2017-05-13 19:02:54.401187] I [MSGID: 108026] >> [afr-self-heal-common.c:1255:afr_log_selfheal] 0-storage2-replicate-2: >> Completed metadata selfheal on fe651475-226e-42a3-be2d-751d4f58e383. >> sources=[0] 1 sinks= >> [2017-05-13 19:02:57.830019] I [MSGID: 109066] >> [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming >> /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. >> 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.par2.tmp >> (hash=storage2-readdir-ahead-2/cache=storage2-readdir-ahead-2) => >> /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. >> 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.par2 >> (hash=storage2-readdir-ahead-0/cache=) >> >> [2017-05-13 19:08:22.014899] I [MSGID: 109066] >> [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming >> /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. >> 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.srr.tmp >> (hash=storage2-readdir-ahead-1/cache=storage2-readdir-ahead-1) => >> /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. >> 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.srr >> (hash=storage2-readdir-ahead-1/cache=) >> [2017-05-13 19:08:22.463840] I [MSGID: 109066] >> [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming >> /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. >> 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r04.tmp >> (hash=storage2-readdir-ahead-2/cache=storage2-readdir-ahead-2) => >> /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. >> 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r04 >> (hash=storage2-readdir-ahead-0/cache=) >> [2017-05-13 19:08:22.769542] I [MSGID:
Re: [Gluster-users] Reliability issues with Gluster 3.10 and shard
I resolved this with the following settings, particularly disabling features.ctr-enabled olume Name: storage2 Type: Distributed-Replicate Volume ID: adaabca5-25ed-4e7f-ae86-2f20fc0143a8 Status: Started Snapshot Count: 0 Number of Bricks: 3 x (2 + 1) = 9 Transport-type: tcp Bricks: Brick1: fd00:0:0:3::6:/mnt/gluster/storage/brick0/glusterfs2 Brick2: fd00:0:0:3::8:/mnt/gluster/storage/brick0/glusterfs2 Brick3: fd00:0:0:3::10:/mnt/gluster/storage/brick0/glusterfs (arbiter) Brick4: fd00:0:0:3::6:/mnt/gluster/storage/brick1/glusterfs2 Brick5: fd00:0:0:3::8:/mnt/gluster/storage/brick1/glusterfs2 Brick6: fd00:0:0:3::10:/mnt/gluster/storage/brick1/glusterfs (arbiter) Brick7: fd00:0:0:3::6:/mnt/gluster/storage/brick2/glusterfs2 Brick8: fd00:0:0:3::8:/mnt/gluster/storage/brick2/glusterfs2 Brick9: fd00:0:0:3::10:/mnt/gluster/storage/brick2/glusterfs (arbiter) Options Reconfigured: performance.write-behind-window-size: 4MB performance.cache-invalidation: on transport.keepalive: on performance.write-behind: on performance.read-ahead: on performance.io-cache: on performance.stat-prefetch: on performance.open-behind: on cluster.use-compound-fops: on performance.cache-ima-xattrs: on features.cache-invalidation: on client.event-threads: 4 cluster.data-self-heal-algorithm: full performance.client-io-threads: on server.event-threads: 4 performance.quick-read: on features.scrub: Active features.bitrot: on features.shard: on transport.address-family: inet6 nfs.disable: on server.allow-insecure: on user.cifs: off cluster.quorum-type: auto cluster.server-quorum-type: server cluster.tier-compact: on diagnostics.brick-log-level: WARNING diagnostics.client-log-level: WARNING cluster.self-heal-daemon: enable performance.cache-samba-metadata: on cluster.brick-multiplex: off cluster.enable-shared-storage: enable nfs-ganesha: enable -ben On Sat, May 13, 2017 at 12:20 PM, Benjamin Kingstonwrote: > Hers's some log entries from nfs-ganesha gfapi > > [2017-05-13 19:02:54.105936] E [MSGID: 133010] > [shard.c:1706:shard_common_lookup_shards_cbk] > 0-storage2-shard: Lookup on shard 11 failed. Base file gfid = > 1494c083-a618-4eba-80a0-147e656dd9d0 [Input/output error] > [2017-05-13 19:02:54.106176] E [MSGID: 133010] > [shard.c:1706:shard_common_lookup_shards_cbk] > 0-storage2-shard: Lookup on shard 2 failed. Base file gfid = > 1494c083-a618-4eba-80a0-147e656dd9d0 [Input/output error] > [2017-05-13 19:02:54.106288] E [MSGID: 133010] > [shard.c:1706:shard_common_lookup_shards_cbk] > 0-storage2-shard: Lookup on shard 1 failed. Base file gfid = > 1494c083-a618-4eba-80a0-147e656dd9d0 [Input/output error] > [2017-05-13 19:02:54.384922] I [MSGID: 108026] > [afr-self-heal-metadata.c:52:__afr_selfheal_metadata_do] > 0-storage2-replicate-2: performing metadata selfheal on > fe651475-226e-42a3-be2d-751d4f58e383 > [2017-05-13 19:02:54.385894] W [MSGID: 114031] > [client-rpc-fops.c:2258:client3_3_setattr_cbk] > 0-storage2-client-8: remote operation failed [Operation not permitted] > [2017-05-13 19:02:54.401187] I [MSGID: 108026] > [afr-self-heal-common.c:1255:afr_log_selfheal] 0-storage2-replicate-2: > Completed metadata selfheal on fe651475-226e-42a3-be2d-751d4f58e383. > sources=[0] 1 sinks= > [2017-05-13 19:02:57.830019] I [MSGID: 109066] > [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.par2.tmp > (hash=storage2-readdir-ahead-2/cache=storage2-readdir-ahead-2) => > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.par2 > (hash=storage2-readdir-ahead-0/cache=) > > [2017-05-13 19:08:22.014899] I [MSGID: 109066] > [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.srr.tmp > (hash=storage2-readdir-ahead-1/cache=storage2-readdir-ahead-1) => > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.srr > (hash=storage2-readdir-ahead-1/cache=) > [2017-05-13 19:08:22.463840] I [MSGID: 109066] > [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r04.tmp > (hash=storage2-readdir-ahead-2/cache=storage2-readdir-ahead-2) => > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r04 > (hash=storage2-readdir-ahead-0/cache=) > [2017-05-13 19:08:22.769542] I [MSGID: 109066] > [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r01.tmp > (hash=storage2-readdir-ahead-2/cache=storage2-readdir-ahead-2) => > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. >
Re: [Gluster-users] Reliability issues with Gluster 3.10 and shard
Shard translator is currently supported only for VM image store workload. -Krutika On Sun, May 14, 2017 at 12:50 AM, Benjamin Kingstonwrote: > Hers's some log entries from nfs-ganesha gfapi > > [2017-05-13 19:02:54.105936] E [MSGID: 133010] > [shard.c:1706:shard_common_lookup_shards_cbk] > 0-storage2-shard: Lookup on shard 11 failed. Base file gfid = > 1494c083-a618-4eba-80a0-147e656dd9d0 [Input/output error] > [2017-05-13 19:02:54.106176] E [MSGID: 133010] > [shard.c:1706:shard_common_lookup_shards_cbk] > 0-storage2-shard: Lookup on shard 2 failed. Base file gfid = > 1494c083-a618-4eba-80a0-147e656dd9d0 [Input/output error] > [2017-05-13 19:02:54.106288] E [MSGID: 133010] > [shard.c:1706:shard_common_lookup_shards_cbk] > 0-storage2-shard: Lookup on shard 1 failed. Base file gfid = > 1494c083-a618-4eba-80a0-147e656dd9d0 [Input/output error] > [2017-05-13 19:02:54.384922] I [MSGID: 108026] > [afr-self-heal-metadata.c:52:__afr_selfheal_metadata_do] > 0-storage2-replicate-2: performing metadata selfheal on > fe651475-226e-42a3-be2d-751d4f58e383 > [2017-05-13 19:02:54.385894] W [MSGID: 114031] > [client-rpc-fops.c:2258:client3_3_setattr_cbk] > 0-storage2-client-8: remote operation failed [Operation not permitted] > [2017-05-13 19:02:54.401187] I [MSGID: 108026] > [afr-self-heal-common.c:1255:afr_log_selfheal] 0-storage2-replicate-2: > Completed metadata selfheal on fe651475-226e-42a3-be2d-751d4f58e383. > sources=[0] 1 sinks= > [2017-05-13 19:02:57.830019] I [MSGID: 109066] > [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.par2.tmp > (hash=storage2-readdir-ahead-2/cache=storage2-readdir-ahead-2) => > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.par2 > (hash=storage2-readdir-ahead-0/cache=) > > [2017-05-13 19:08:22.014899] I [MSGID: 109066] > [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.srr.tmp > (hash=storage2-readdir-ahead-1/cache=storage2-readdir-ahead-1) => > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.srr > (hash=storage2-readdir-ahead-1/cache=) > [2017-05-13 19:08:22.463840] I [MSGID: 109066] > [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r04.tmp > (hash=storage2-readdir-ahead-2/cache=storage2-readdir-ahead-2) => > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r04 > (hash=storage2-readdir-ahead-0/cache=) > [2017-05-13 19:08:22.769542] I [MSGID: 109066] > [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r01.tmp > (hash=storage2-readdir-ahead-2/cache=storage2-readdir-ahead-2) => > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r01 > (hash=storage2-readdir-ahead-0/cache=) > [2017-05-13 19:08:23.141069] I [MSGID: 109066] > [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.nfo.tmp > (hash=storage2-readdir-ahead-1/cache=storage2-readdir-ahead-1) => > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.nfo > (hash=storage2-readdir-ahead-0/cache=) > [2017-05-13 19:08:23.468554] I [MSGID: 109066] > [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r00.tmp > (hash=storage2-readdir-ahead-0/cache=storage2-readdir-ahead-0) => > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r00 > (hash=storage2-readdir-ahead-2/cache=) > [2017-05-13 19:08:23.671753] I [MSGID: 109066] > [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.sfv.tmp > (hash=storage2-readdir-ahead-2/cache=storage2-readdir-ahead-2) => > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.sfv > (hash=storage2-readdir-ahead-2/cache=) > [2017-05-13 19:08:23.812152] I [MSGID: 109066] > [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming > /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05. > 720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r11.tmp >
Re: [Gluster-users] Reliability issues with Gluster 3.10 and shard
Hers's some log entries from nfs-ganesha gfapi [2017-05-13 19:02:54.105936] E [MSGID: 133010] [shard.c:1706:shard_common_lookup_shards_cbk] 0-storage2-shard: Lookup on shard 11 failed. Base file gfid = 1494c083-a618-4eba-80a0-147e656dd9d0 [Input/output error] [2017-05-13 19:02:54.106176] E [MSGID: 133010] [shard.c:1706:shard_common_lookup_shards_cbk] 0-storage2-shard: Lookup on shard 2 failed. Base file gfid = 1494c083-a618-4eba-80a0-147e656dd9d0 [Input/output error] [2017-05-13 19:02:54.106288] E [MSGID: 133010] [shard.c:1706:shard_common_lookup_shards_cbk] 0-storage2-shard: Lookup on shard 1 failed. Base file gfid = 1494c083-a618-4eba-80a0-147e656dd9d0 [Input/output error] [2017-05-13 19:02:54.384922] I [MSGID: 108026] [afr-self-heal-metadata.c:52:__afr_selfheal_metadata_do] 0-storage2-replicate-2: performing metadata selfheal on fe651475-226e-42a3-be2d-751d4f58e383 [2017-05-13 19:02:54.385894] W [MSGID: 114031] [client-rpc-fops.c:2258:client3_3_setattr_cbk] 0-storage2-client-8: remote operation failed [Operation not permitted] [2017-05-13 19:02:54.401187] I [MSGID: 108026] [afr-self-heal-common.c:1255:afr_log_selfheal] 0-storage2-replicate-2: Completed metadata selfheal on fe651475-226e-42a3-be2d-751d4f58e383. sources=[0] 1 sinks= [2017-05-13 19:02:57.830019] I [MSGID: 109066] [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.par2.tmp (hash=storage2-readdir-ahead-2/cache=storage2-readdir-ahead-2) => /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.par2 (hash=storage2-readdir-ahead-0/cache=) [2017-05-13 19:08:22.014899] I [MSGID: 109066] [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.srr.tmp (hash=storage2-readdir-ahead-1/cache=storage2-readdir-ahead-1) => /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.srr (hash=storage2-readdir-ahead-1/cache=) [2017-05-13 19:08:22.463840] I [MSGID: 109066] [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r04.tmp (hash=storage2-readdir-ahead-2/cache=storage2-readdir-ahead-2) => /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r04 (hash=storage2-readdir-ahead-0/cache=) [2017-05-13 19:08:22.769542] I [MSGID: 109066] [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r01.tmp (hash=storage2-readdir-ahead-2/cache=storage2-readdir-ahead-2) => /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r01 (hash=storage2-readdir-ahead-0/cache=) [2017-05-13 19:08:23.141069] I [MSGID: 109066] [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.nfo.tmp (hash=storage2-readdir-ahead-1/cache=storage2-readdir-ahead-1) => /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.nfo (hash=storage2-readdir-ahead-0/cache=) [2017-05-13 19:08:23.468554] I [MSGID: 109066] [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r00.tmp (hash=storage2-readdir-ahead-0/cache=storage2-readdir-ahead-0) => /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r00 (hash=storage2-readdir-ahead-2/cache=) [2017-05-13 19:08:23.671753] I [MSGID: 109066] [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.sfv.tmp (hash=storage2-readdir-ahead-2/cache=storage2-readdir-ahead-2) => /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.sfv (hash=storage2-readdir-ahead-2/cache=) [2017-05-13 19:08:23.812152] I [MSGID: 109066] [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r11.tmp (hash=storage2-readdir-ahead-0/cache=storage2-readdir-ahead-0) => /content/Downloads/incomplete/usenet/Attack.on.Titan.S02E05.720p.WEB.x264-ANiURL.#27/aniurl-aot.s02e05.720p.web.r11 (hash=storage2-readdir-ahead-1/cache=) [2017-05-13 19:08:24.244584] I [MSGID: 109066] [dht-rename.c:1608:dht_rename] 0-storage2-dht: renaming
[Gluster-users] Reliability issues with Gluster 3.10 and shard
Hello all, I'm trying to take advantage of the shard xlator, however I've found it causes a lot of issues that I hope is easily resolvable 1) large file operations work well (copy file from folder a to folder b 2) seek operations and list operations frequently fail (ls directory, read bytes xyz at offset 235567) Turning off the shard feature resolves this issue for new files created in the volume. mounted using the gluster fuse mount here's my volume settings, please let me know if there's some changes I can make. Option Value -- - cluster.lookup-unhashed on cluster.lookup-optimize on cluster.min-free-disk 10% cluster.min-free-inodes 5% cluster.rebalance-stats off cluster.subvols-per-directory (null) cluster.readdir-optimize off cluster.rsync-hash-regex (null) cluster.extra-hash-regex (null) cluster.dht-xattr-name trusted.glusterfs.dht cluster.randomize-hash-range-by-gfid off cluster.rebal-throttle normal cluster.lock-migration off cluster.local-volume-name (null) cluster.weighted-rebalance on cluster.switch-pattern (null) cluster.entry-change-log on cluster.read-subvolume (null) cluster.read-subvolume-index -1 cluster.read-hash-mode 1 cluster.background-self-heal-count 8 cluster.metadata-self-heal on cluster.data-self-heal on cluster.entry-self-heal on cluster.self-heal-daemon on cluster.heal-timeout 600 cluster.self-heal-window-size 1 cluster.data-change-log on cluster.metadata-change-log on cluster.data-self-heal-algorithm diff cluster.eager-lock enable disperse.eager-lock on cluster.quorum-type auto cluster.quorum-count (null) cluster.choose-local on cluster.self-heal-readdir-size 1KB cluster.post-op-delay-secs 1 cluster.ensure-durability on cluster.consistent-metadata no cluster.heal-wait-queue-length 128 cluster.favorite-child-policy none cluster.stripe-block-size 128KB cluster.stripe-coalesce true diagnostics.latency-measurement off diagnostics.dump-fd-stats off diagnostics.count-fop-hits off diagnostics.brick-log-level INFO diagnostics.client-log-level INFO diagnostics.brick-sys-log-level CRITICAL diagnostics.client-sys-log-level CRITICAL diagnostics.brick-logger (null) diagnostics.client-logger (null) diagnostics.brick-log-format (null) diagnostics.client-log-format (null) diagnostics.brick-log-buf-size 5 diagnostics.client-log-buf-size 5 diagnostics.brick-log-flush-timeout 120 diagnostics.client-log-flush-timeout 120 diagnostics.stats-dump-interval 0 diagnostics.fop-sample-interval 0 diagnostics.fop-sample-buf-size 65535 diagnostics.stats-dnscache-ttl-sec 86400 performance.cache-max-file-size 0 performance.cache-min-file-size 0 performance.cache-refresh-timeout 1 performance.cache-priority performance.cache-size 1GB performance.io-thread-count 64 performance.high-prio-threads 16 performance.normal-prio-threads 16 performance.low-prio-threads 32 performance.least-prio-threads 1 performance.enable-least-priority on performance.cache-size 1GB performance.flush-behind on performance.nfs.flush-behind on performance.write-behind-window-size 2GB performance.resync-failed-syncs-after-fsyncoff performance.nfs.write-behind-window-size1MB performance.strict-o-direct off performance.nfs.strict-o-direct off performance.strict-write-ordering off performance.nfs.strict-write-ordering off performance.lazy-open yes performance.read-after-open no performance.read-ahead-page-count 4 performance.md-cache-timeout 1 performance.cache-swift-metadata true performance.cache-samba-metadata false performance.cache-capability-xattrs true performance.cache-ima-xattrs on features.encryption off encryption.master-key (null) encryption.data-key-size 256 encryption.block-size 4096 network.frame-timeout 1800 network.ping-timeout 42 network.tcp-window-size (null) features.lock-heal off features.grace-timeout 10 network.remote-dio disable client.event-threads 3 network.ping-timeout 42 network.tcp-window-size (null) network.inode-lru-limit 9 auth.allow * auth.reject (null) transport.keepalive on server.allow-insecure on server.root-squash off server.anonuid 65534 server.anongid 65534 server.statedump-path /var/run/gluster server.outstanding-rpc-limit 64 features.lock-heal off features.grace-timeout 10 server.ssl (null) auth.ssl-allow * server.manage-gids off server.dynamic-auth on client.send-gids on server.gid-timeout 300 server.own-thread (null) server.event-threads 3 ssl.own-cert (null) ssl.private-key (null) ssl.ca-list (null) ssl.crl-path (null) ssl.certificate-depth (null) ssl.cipher-list (null) ssl.dh-param (null) ssl.ec-curve (null) transport.address-family inet6 performance.write-behind on performance.read-ahead off performance.readdir-ahead on performance.io-cache on performance.quick-read off performance.open-behind on performance.stat-prefetch on performance.client-io-threads on performance.nfs.write-behind on performance.nfs.read-ahead off performance.nfs.io-cache off performance.nfs.quick-read off performance.nfs.stat-prefetch off performance.nfs.io-threads off performance.force-readdirp true