Hi Nithya, thank you very much for your support and sorry for the late. Below you can find the output of “gluster volume info tier2” command and the gluster software stack version:
gluster volume info Volume Name: tier2 Type: Distributed-Disperse Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c Status: Started Snapshot Count: 0 Number of Bricks: 6 x (4 + 2) = 36 Transport-type: tcp Bricks: Brick1: s01-stg:/gluster/mnt1/brick Brick2: s02-stg:/gluster/mnt1/brick Brick3: s03-stg:/gluster/mnt1/brick Brick4: s01-stg:/gluster/mnt2/brick Brick5: s02-stg:/gluster/mnt2/brick Brick6: s03-stg:/gluster/mnt2/brick Brick7: s01-stg:/gluster/mnt3/brick Brick8: s02-stg:/gluster/mnt3/brick Brick9: s03-stg:/gluster/mnt3/brick Brick10: s01-stg:/gluster/mnt4/brick Brick11: s02-stg:/gluster/mnt4/brick Brick12: s03-stg:/gluster/mnt4/brick Brick13: s01-stg:/gluster/mnt5/brick Brick14: s02-stg:/gluster/mnt5/brick Brick15: s03-stg:/gluster/mnt5/brick Brick16: s01-stg:/gluster/mnt6/brick Brick17: s02-stg:/gluster/mnt6/brick Brick18: s03-stg:/gluster/mnt6/brick Brick19: s01-stg:/gluster/mnt7/brick Brick20: s02-stg:/gluster/mnt7/brick Brick21: s03-stg:/gluster/mnt7/brick Brick22: s01-stg:/gluster/mnt8/brick Brick23: s02-stg:/gluster/mnt8/brick Brick24: s03-stg:/gluster/mnt8/brick Brick25: s01-stg:/gluster/mnt9/brick Brick26: s02-stg:/gluster/mnt9/brick Brick27: s03-stg:/gluster/mnt9/brick Brick28: s01-stg:/gluster/mnt10/brick Brick29: s02-stg:/gluster/mnt10/brick Brick30: s03-stg:/gluster/mnt10/brick Brick31: s01-stg:/gluster/mnt11/brick Brick32: s02-stg:/gluster/mnt11/brick Brick33: s03-stg:/gluster/mnt11/brick Brick34: s01-stg:/gluster/mnt12/brick Brick35: s02-stg:/gluster/mnt12/brick Brick36: s03-stg:/gluster/mnt12/brick Options Reconfigured: features.scrub: Active features.bitrot: on features.inode-quota: on features.quota: on performance.client-io-threads: on cluster.min-free-disk: 10 cluster.quorum-type: auto transport.address-family: inet nfs.disable: on server.event-threads: 4 client.event-threads: 4 cluster.lookup-optimize: on performance.readdir-ahead: on performance.parallel-readdir: off cluster.readdir-optimize: on features.cache-invalidation: on features.cache-invalidation-timeout: 600 performance.stat-prefetch: on performance.cache-invalidation: on performance.md-cache-timeout: 600 network.inode-lru-limit: 50000 performance.io-cache: off disperse.cpu-extensions: auto performance.io-thread-count: 16 features.quota-deem-statfs: on features.default-soft-limit: 90 cluster.server-quorum-type: server cluster.brick-multiplex: on cluster.server-quorum-ratio: 51% [root@s01 ~]# rpm -qa|grep gluster python2-gluster-3.10.5-1.el7.x86_64 glusterfs-geo-replication-3.10.5-1.el7.x86_64 centos-release-gluster310-1.0-1.el7.centos.noarch glusterfs-server-3.10.5-1.el7.x86_64 glusterfs-libs-3.10.5-1.el7.x86_64 glusterfs-api-3.10.5-1.el7.x86_64 vdsm-gluster-4.19.31-1.el7.centos.noarch glusterfs-3.10.5-1.el7.x86_64 gluster-nagios-common-1.1.0-0.el7.centos.noarch glusterfs-cli-3.10.5-1.el7.x86_64 glusterfs-client-xlators-3.10.5-1.el7.x86_64 gluster-nagios-addons-1.1.0-0.el7.centos.x86_64 glusterfs-fuse-3.10.5-1.el7.x86_64 libvirt-daemon-driver-storage-gluster-3.2.0-14.el7_4.3.x86_64 Many thanks, Mauro > Il giorno 29 dic 2017, alle ore 04:59, Nithya Balachandran > <[email protected]> ha scritto: > > Hi Mauro, > > What version of Gluster are you running and what is your volume configuration? > > IIRC, this was seen because of mismatches in the ctime returned to the > client. I don't think there were issues with the files but I will leave it to > Ravi and Raghavendra to comment. > > > Regards, > Nithya > > > On 29 December 2017 at 04:10, Mauro Tridici <[email protected] > <mailto:[email protected]>> wrote: > > Hi All, > > anyone had the same experience? > Could you provide me some information about this error? > It happens only on GlusterFS file system. > > Thank you, > Mauro > >> Il giorno 20 dic 2017, alle ore 16:57, Mauro Tridici <[email protected] >> <mailto:[email protected]>> ha scritto: >> >> >> Dear Users, >> >> I’m experiencing a random problem ( "file changed as we read it” error) >> during tar files creation on a distributed dispersed Gluster file system. >> >> The tar files seem to be created correctly, but I can see a lot of message >> similar to the following ones: >> >> tar: ./year1990/lffd1990050706p.nc <http://lffd1990050706p.nc/>.gz: file >> changed as we read it >> tar: ./year1990/lffd1990052106p.nc <http://lffd1990052106p.nc/>.gz: file >> changed as we read it >> tar: ./year1990/lffd1990052412p.nc <http://lffd1990052412p.nc/>.gz: file >> changed as we read it >> tar: ./year1990/lffd1990091018.nc <http://lffd1990091018.nc/>.gz: file >> changed as we read it >> tar: ./year1990/lffd1990092300p.nc <http://lffd1990092300p.nc/>.gz: file >> changed as we read it >> tar: ./year1990/lffd1990092706p.nc <http://lffd1990092706p.nc/>.gz: file >> changed as we read it >> tar: ./year1990/lffd1990100312p.nc <http://lffd1990100312p.nc/>.gz: file >> changed as we read it >> tar: ./year1990/lffd1990100412.nc <http://lffd1990100412.nc/>.gz: file >> changed as we read it >> tar: ./year1991/lffd1991012106.nc <http://lffd1991012106.nc/>.gz: file >> changed as we read it >> tar: ./year1991/lffd1991010918.nc <http://lffd1991010918.nc/>.gz: file >> changed as we read it >> tar: ./year1991/lffd1991011400.nc <http://lffd1991011400.nc/>.gz: file >> changed as we read it >> >> I’m executing the tar command on a CentOS 6.2 operating system based server: >> it is a gluster native client. >> >> You can find below some basic info about the gluster client: >> >> [root@athena# rpm -qa|grep gluster >> glusterfs-3.10.5-1.el6.x86_64 >> centos-release-gluster310-1.0-1.el6.centos.noarch >> glusterfs-client-xlators-3.10.5-1.el6.x86_64 >> glusterfs-fuse-3.10.5-1.el6.x86_64 >> glusterfs-libs-3.10.5-1.el6.x86_64 >> >> Can I consider them as a false positive or the created tar files will suffer >> of inconsistence? >> Is it a tar command problem or a gluster problem? >> >> Could someone help me to resolve this issue? >> >> Thank you very much, >> Mauro > > > > > > _______________________________________________ > Gluster-users mailing list > [email protected] <mailto:[email protected]> > http://lists.gluster.org/mailman/listinfo/gluster-users > <http://lists.gluster.org/mailman/listinfo/gluster-users> > ------------------------- Mauro Tridici Fondazione CMCC CMCC Supercomputing Center presso Complesso Ecotekne - Università del Salento - Strada Prov.le Lecce - Monteroni sn 73100 Lecce IT http://www.cmcc.it mobile: (+39) 327 5630841 email: [email protected]
_______________________________________________ Gluster-users mailing list [email protected] http://lists.gluster.org/mailman/listinfo/gluster-users
