OK. I am not sure what it is that we're doing differently. I tried the steps you shared and here's what I got:
[root@dhcp35-215 bricks]# gluster volume info Volume Name: rep Type: Replicate Volume ID: 3fd45a4b-0d02-4a44-b74a-41592d48e102 Status: Started Number of Bricks: 1 x 3 = 3 Transport-type: tcp Bricks: Brick1: kdhananjay:/bricks/1 Brick2: kdhananjay:/bricks/2 Brick3: kdhananjay:/bricks/3 Options Reconfigured: performance.strict-write-ordering: on features.shard: on features.shard-block-size: 512MB cluster.quorum-type: auto client.event-threads: 4 server.event-threads: 4 cluster.self-heal-window-size: 256 performance.write-behind: on nfs.enable-ino32: on nfs.addr-namelookup: off nfs.disable: on performance.cache-refresh-timeout: 4 performance.cache-size: 1GB performance.write-behind-window-size: 128MB performance.io-thread-count: 32 performance.readdir-ahead: on [root@dhcp35-215 mnt]# gluster volume set rep strict-write-ordering on volume set: success [root@dhcp35-215 mnt]# dd if=/dev/sda of=test.bin bs=1MB count=8192 8192+0 records in 8192+0 records out 8192000000 bytes (8.2 GB) copied, 133.754 s, 61.2 MB/s [root@dhcp35-215 mnt]# ls -l total 8000000 -rw-r--r--. 1 root root 8192000000 Nov 5 16:40 test.bin [root@dhcp35-215 mnt]# ls -lh total 7.7G -rw-r--r--. 1 root root 7.7G Nov 5 16:40 test.bin [root@dhcp35-215 mnt]# du test.bin 8000000 test.bin [root@dhcp35-215 bricks]# du /bricks/1/.shard/ 7475780 /bricks/1/.shard/ [root@dhcp35-215 bricks]# du /bricks/1/ .glusterfs/ .shard/ test.bin .trashcan/ [root@dhcp35-215 bricks]# du /bricks/1/test.bin 524292 /bricks/1/test.bin Just to be sure, did you rerun the test on the already broken file (test.bin) which was written to when strict-write-ordering had been off? Or did you try the new test with strict-write-ordering on a brand new file? -Krutika ----- Original Message ----- > From: "Lindsay Mathieson" <[email protected]> > To: "Krutika Dhananjay" <[email protected]> > Cc: "gluster-users" <[email protected]> > Sent: Thursday, November 5, 2015 3:04:51 AM > Subject: Re: [Gluster-users] Shard file size (gluster 3.7.5) > On 5 November 2015 at 01:09, Krutika Dhananjay < [email protected] > wrote: > > Ah! It's the same issue. Just saw your volume info output. Enabling > > strict-write-ordering should ensure both size and disk usage are accurate. > > Tested it - nope :( Size s accurate (27746172928 bytes), but disk usage is > wildly inaccurate (698787). > I have compression disabled on the underlying storage now. > -- > Lindsay
_______________________________________________ Gluster-users mailing list [email protected] http://www.gluster.org/mailman/listinfo/gluster-users
