Yes please, could you file a bug against glusterfs for this issue?
-Krutika On Wed, Jul 27, 2016 at 1:39 AM, David Gossage <dgoss...@carouselchecks.com> wrote: > Has a bug report been filed for this issue or should l I create one with > the logs and results provided so far? > > *David Gossage* > *Carousel Checks Inc. | System Administrator* > *Office* 708.613.2284 > > On Fri, Jul 22, 2016 at 12:53 PM, David Gossage < > dgoss...@carouselchecks.com> wrote: > >> >> >> >> On Fri, Jul 22, 2016 at 9:37 AM, Vijay Bellur <vbel...@redhat.com> wrote: >> >>> On Fri, Jul 22, 2016 at 10:03 AM, Samuli Heinonen <samp...@neutraali.net> >>> wrote: >>> > Here is a quick way how to test this: >>> > GlusterFS 3.7.13 volume with default settings with brick on ZFS >>> dataset. gluster-test1 is server and gluster-test2 is client mounting with >>> FUSE. >>> > >>> > Writing file with oflag=direct is not ok: >>> > [root@gluster-test2 gluster]# dd if=/dev/zero of=file oflag=direct >>> count=1 bs=1024000 >>> > dd: failed to open ‘file’: Invalid argument >>> > >>> > Enable network.remote-dio on Gluster Volume: >>> > [root@gluster-test1 gluster]# gluster volume set gluster >>> network.remote-dio enable >>> > volume set: success >>> > >>> > Writing small file with oflag=direct is ok: >>> > [root@gluster-test2 gluster]# dd if=/dev/zero of=file oflag=direct >>> count=1 bs=1024000 >>> > 1+0 records in >>> > 1+0 records out >>> > 1024000 bytes (1.0 MB) copied, 0.0103793 s, 98.7 MB/s >>> > >>> > Writing bigger file with oflag=direct is ok: >>> > [root@gluster-test2 gluster]# dd if=/dev/zero of=file3 oflag=direct >>> count=100 bs=1M >>> > 100+0 records in >>> > 100+0 records out >>> > 104857600 bytes (105 MB) copied, 1.10583 s, 94.8 MB/s >>> > >>> > Enable Sharding on Gluster Volume: >>> > [root@gluster-test1 gluster]# gluster volume set gluster >>> features.shard enable >>> > volume set: success >>> > >>> > Writing small file with oflag=direct is ok: >>> > [root@gluster-test2 gluster]# dd if=/dev/zero of=file3 oflag=direct >>> count=1 bs=1M >>> > 1+0 records in >>> > 1+0 records out >>> > 1048576 bytes (1.0 MB) copied, 0.0115247 s, 91.0 MB/s >>> > >>> > Writing bigger file with oflag=direct is not ok: >>> > [root@gluster-test2 gluster]# dd if=/dev/zero of=file3 oflag=direct >>> count=100 bs=1M >>> > dd: error writing ‘file3’: Operation not permitted >>> > dd: closing output file ‘file3’: Operation not permitted >>> > >>> >>> >>> Thank you for these tests! would it be possible to share the brick and >>> client logs? >>> >> >> Not sure if his tests are same as my setup but here is what I end up with >> >> Volume Name: glustershard >> Type: Replicate >> Volume ID: 0cc4efb6-3836-4caa-b24a-b3afb6e407c3 >> Status: Started >> Number of Bricks: 1 x 3 = 3 >> Transport-type: tcp >> Bricks: >> Brick1: 192.168.71.10:/gluster1/shard1/1 >> Brick2: 192.168.71.11:/gluster1/shard2/1 >> Brick3: 192.168.71.12:/gluster1/shard3/1 >> Options Reconfigured: >> features.shard-block-size: 64MB >> features.shard: on >> server.allow-insecure: on >> storage.owner-uid: 36 >> storage.owner-gid: 36 >> cluster.server-quorum-type: server >> cluster.quorum-type: auto >> network.remote-dio: enable >> cluster.eager-lock: enable >> performance.stat-prefetch: off >> performance.io-cache: off >> performance.quick-read: off >> cluster.self-heal-window-size: 1024 >> cluster.background-self-heal-count: 16 >> nfs.enable-ino32: off >> nfs.addr-namelookup: off >> nfs.disable: on >> performance.read-ahead: off >> performance.readdir-ahead: on >> >> >> >> dd if=/dev/zero >> of=/rhev/data-center/mnt/glusterSD/192.168.71.11\:_glustershard/ >> oflag=direct count=100 bs=1M >> 81e19cd3-ae45-449c-b716-ec3e4ad4c2f0/ __DIRECT_IO_TEST__ >> .trashcan/ >> [root@ccengine2 ~]# dd if=/dev/zero of=/rhev/data-center/mnt/glusterSD/ >> 192.168.71.11\:_glustershard/81e19cd3-ae45-449c-b716-ec3e4ad4c2f0/images/test >> oflag=direct count=100 bs=1M >> dd: error writing >> ‘/rhev/data-center/mnt/glusterSD/192.168.71.11:_glustershard/81e19cd3-ae45-449c-b716-ec3e4ad4c2f0/images/test’: >> Operation not permitted >> >> creates the 64M file in expected location then the shard is 0 >> >> # file: gluster1/shard1/1/81e19cd3-ae45-449c-b716-ec3e4ad4c2f0/images/test >> >> security.selinux=0x73797374656d5f753a6f626a6563745f723a756e6c6162656c65645f743a733000 >> trusted.afr.dirty=0x000000000000000000000000 >> trusted.bit-rot.version=0x0200000000000000579231f3000e16e7 >> trusted.gfid=0xec6de302b35f427985639ca3e25d9df0 >> trusted.glusterfs.shard.block-size=0x0000000004000000 >> >> trusted.glusterfs.shard.file-size=0x0000000004000000000000000000000000000000000000010000000000000000 >> >> >> # file: gluster1/shard1/1/.shard/ec6de302-b35f-4279-8563-9ca3e25d9df0.1 >> >> security.selinux=0x73797374656d5f753a6f626a6563745f723a756e6c6162656c65645f743a733000 >> trusted.afr.dirty=0x000000000000000000000000 >> trusted.gfid=0x2bfd3cc8a727489b9a0474241548fe80 >> >> >>> Regards, >>> Vijay >>> _______________________________________________ >>> Gluster-users mailing list >>> gluster-us...@gluster.org >>> http://www.gluster.org/mailman/listinfo/gluster-users >> >> >> > > _______________________________________________ > Gluster-devel mailing list > Gluster-devel@gluster.org > http://www.gluster.org/mailman/listinfo/gluster-devel >
_______________________________________________ Gluster-devel mailing list Gluster-devel@gluster.org http://www.gluster.org/mailman/listinfo/gluster-devel