Hi All, any news about this issue? Can I ignore this kind of error message or I have to do something to correct it?
Thank you in advance and sorry for my insistence. Regards, Mauro > Il giorno 29 dic 2017, alle ore 11:45, Mauro Tridici <[email protected]> > ha scritto: > > > Hi Nithya, > > thank you very much for your support and sorry for the late. > Below you can find the output of “gluster volume info tier2” command and the > gluster software stack version: > > gluster volume info > > Volume Name: tier2 > Type: Distributed-Disperse > Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c > Status: Started > Snapshot Count: 0 > Number of Bricks: 6 x (4 + 2) = 36 > Transport-type: tcp > Bricks: > Brick1: s01-stg:/gluster/mnt1/brick > Brick2: s02-stg:/gluster/mnt1/brick > Brick3: s03-stg:/gluster/mnt1/brick > Brick4: s01-stg:/gluster/mnt2/brick > Brick5: s02-stg:/gluster/mnt2/brick > Brick6: s03-stg:/gluster/mnt2/brick > Brick7: s01-stg:/gluster/mnt3/brick > Brick8: s02-stg:/gluster/mnt3/brick > Brick9: s03-stg:/gluster/mnt3/brick > Brick10: s01-stg:/gluster/mnt4/brick > Brick11: s02-stg:/gluster/mnt4/brick > Brick12: s03-stg:/gluster/mnt4/brick > Brick13: s01-stg:/gluster/mnt5/brick > Brick14: s02-stg:/gluster/mnt5/brick > Brick15: s03-stg:/gluster/mnt5/brick > Brick16: s01-stg:/gluster/mnt6/brick > Brick17: s02-stg:/gluster/mnt6/brick > Brick18: s03-stg:/gluster/mnt6/brick > Brick19: s01-stg:/gluster/mnt7/brick > Brick20: s02-stg:/gluster/mnt7/brick > Brick21: s03-stg:/gluster/mnt7/brick > Brick22: s01-stg:/gluster/mnt8/brick > Brick23: s02-stg:/gluster/mnt8/brick > Brick24: s03-stg:/gluster/mnt8/brick > Brick25: s01-stg:/gluster/mnt9/brick > Brick26: s02-stg:/gluster/mnt9/brick > Brick27: s03-stg:/gluster/mnt9/brick > Brick28: s01-stg:/gluster/mnt10/brick > Brick29: s02-stg:/gluster/mnt10/brick > Brick30: s03-stg:/gluster/mnt10/brick > Brick31: s01-stg:/gluster/mnt11/brick > Brick32: s02-stg:/gluster/mnt11/brick > Brick33: s03-stg:/gluster/mnt11/brick > Brick34: s01-stg:/gluster/mnt12/brick > Brick35: s02-stg:/gluster/mnt12/brick > Brick36: s03-stg:/gluster/mnt12/brick > Options Reconfigured: > features.scrub: Active > features.bitrot: on > features.inode-quota: on > features.quota: on > performance.client-io-threads: on > cluster.min-free-disk: 10 > cluster.quorum-type: auto > transport.address-family: inet > nfs.disable: on > server.event-threads: 4 > client.event-threads: 4 > cluster.lookup-optimize: on > performance.readdir-ahead: on > performance.parallel-readdir: off > cluster.readdir-optimize: on > features.cache-invalidation: on > features.cache-invalidation-timeout: 600 > performance.stat-prefetch: on > performance.cache-invalidation: on > performance.md-cache-timeout: 600 > network.inode-lru-limit: 50000 > performance.io <http://performance.io/>-cache: off > disperse.cpu-extensions: auto > performance.io <http://performance.io/>-thread-count: 16 > features.quota-deem-statfs: on > features.default-soft-limit: 90 > cluster.server-quorum-type: server > cluster.brick-multiplex: on > cluster.server-quorum-ratio: 51% > > [root@s01 ~]# rpm -qa|grep gluster > python2-gluster-3.10.5-1.el7.x86_64 > glusterfs-geo-replication-3.10.5-1.el7.x86_64 > centos-release-gluster310-1.0-1.el7.centos.noarch > glusterfs-server-3.10.5-1.el7.x86_64 > glusterfs-libs-3.10.5-1.el7.x86_64 > glusterfs-api-3.10.5-1.el7.x86_64 > vdsm-gluster-4.19.31-1.el7.centos.noarch > glusterfs-3.10.5-1.el7.x86_64 > gluster-nagios-common-1.1.0-0.el7.centos.noarch > glusterfs-cli-3.10.5-1.el7.x86_64 > glusterfs-client-xlators-3.10.5-1.el7.x86_64 > gluster-nagios-addons-1.1.0-0.el7.centos.x86_64 > glusterfs-fuse-3.10.5-1.el7.x86_64 > libvirt-daemon-driver-storage-gluster-3.2.0-14.el7_4.3.x86_64 > > Many thanks, > Mauro > >> Il giorno 29 dic 2017, alle ore 04:59, Nithya Balachandran >> <[email protected] <mailto:[email protected]>> ha scritto: >> >> Hi Mauro, >> >> What version of Gluster are you running and what is your volume >> configuration? >> >> IIRC, this was seen because of mismatches in the ctime returned to the >> client. I don't think there were issues with the files but I will leave it >> to Ravi and Raghavendra to comment. >> >> >> Regards, >> Nithya >> >> >> On 29 December 2017 at 04:10, Mauro Tridici <[email protected] >> <mailto:[email protected]>> wrote: >> >> Hi All, >> >> anyone had the same experience? >> Could you provide me some information about this error? >> It happens only on GlusterFS file system. >> >> Thank you, >> Mauro >> >>> Il giorno 20 dic 2017, alle ore 16:57, Mauro Tridici <[email protected] >>> <mailto:[email protected]>> ha scritto: >>> >>> >>> Dear Users, >>> >>> I’m experiencing a random problem ( "file changed as we read it” error) >>> during tar files creation on a distributed dispersed Gluster file system. >>> >>> The tar files seem to be created correctly, but I can see a lot of message >>> similar to the following ones: >>> >>> tar: ./year1990/lffd1990050706p.nc <http://lffd1990050706p.nc/>.gz: file >>> changed as we read it >>> tar: ./year1990/lffd1990052106p.nc <http://lffd1990052106p.nc/>.gz: file >>> changed as we read it >>> tar: ./year1990/lffd1990052412p.nc <http://lffd1990052412p.nc/>.gz: file >>> changed as we read it >>> tar: ./year1990/lffd1990091018.nc <http://lffd1990091018.nc/>.gz: file >>> changed as we read it >>> tar: ./year1990/lffd1990092300p.nc <http://lffd1990092300p.nc/>.gz: file >>> changed as we read it >>> tar: ./year1990/lffd1990092706p.nc <http://lffd1990092706p.nc/>.gz: file >>> changed as we read it >>> tar: ./year1990/lffd1990100312p.nc <http://lffd1990100312p.nc/>.gz: file >>> changed as we read it >>> tar: ./year1990/lffd1990100412.nc <http://lffd1990100412.nc/>.gz: file >>> changed as we read it >>> tar: ./year1991/lffd1991012106.nc <http://lffd1991012106.nc/>.gz: file >>> changed as we read it >>> tar: ./year1991/lffd1991010918.nc <http://lffd1991010918.nc/>.gz: file >>> changed as we read it >>> tar: ./year1991/lffd1991011400.nc <http://lffd1991011400.nc/>.gz: file >>> changed as we read it >>> >>> I’m executing the tar command on a CentOS 6.2 operating system based >>> server: it is a gluster native client. >>> >>> You can find below some basic info about the gluster client: >>> >>> [root@athena# rpm -qa|grep gluster >>> glusterfs-3.10.5-1.el6.x86_64 >>> centos-release-gluster310-1.0-1.el6.centos.noarch >>> glusterfs-client-xlators-3.10.5-1.el6.x86_64 >>> glusterfs-fuse-3.10.5-1.el6.x86_64 >>> glusterfs-libs-3.10.5-1.el6.x86_64 >>> >>> Can I consider them as a false positive or the created tar files will >>> suffer of inconsistence? >>> Is it a tar command problem or a gluster problem? >>> >>> Could someone help me to resolve this issue? >>> >>> Thank you very much, >>> Mauro >> >> >> >> >> >> _______________________________________________ >> Gluster-users mailing list >> [email protected] <mailto:[email protected]> >> http://lists.gluster.org/mailman/listinfo/gluster-users >> <http://lists.gluster.org/mailman/listinfo/gluster-users> >> > >
_______________________________________________ Gluster-users mailing list [email protected] http://lists.gluster.org/mailman/listinfo/gluster-users
