Hi Nithya,

thank you very much for your support and sorry for the late.
Below you can find the output of “gluster volume info tier2” command and the 
gluster software stack version:

gluster volume info
 
Volume Name: tier2
Type: Distributed-Disperse
Volume ID: a28d88c5-3295-4e35-98d4-210b3af9358c
Status: Started
Snapshot Count: 0
Number of Bricks: 6 x (4 + 2) = 36
Transport-type: tcp
Bricks:
Brick1: s01-stg:/gluster/mnt1/brick
Brick2: s02-stg:/gluster/mnt1/brick
Brick3: s03-stg:/gluster/mnt1/brick
Brick4: s01-stg:/gluster/mnt2/brick
Brick5: s02-stg:/gluster/mnt2/brick
Brick6: s03-stg:/gluster/mnt2/brick
Brick7: s01-stg:/gluster/mnt3/brick
Brick8: s02-stg:/gluster/mnt3/brick
Brick9: s03-stg:/gluster/mnt3/brick
Brick10: s01-stg:/gluster/mnt4/brick
Brick11: s02-stg:/gluster/mnt4/brick
Brick12: s03-stg:/gluster/mnt4/brick
Brick13: s01-stg:/gluster/mnt5/brick
Brick14: s02-stg:/gluster/mnt5/brick
Brick15: s03-stg:/gluster/mnt5/brick
Brick16: s01-stg:/gluster/mnt6/brick
Brick17: s02-stg:/gluster/mnt6/brick
Brick18: s03-stg:/gluster/mnt6/brick
Brick19: s01-stg:/gluster/mnt7/brick
Brick20: s02-stg:/gluster/mnt7/brick
Brick21: s03-stg:/gluster/mnt7/brick
Brick22: s01-stg:/gluster/mnt8/brick
Brick23: s02-stg:/gluster/mnt8/brick
Brick24: s03-stg:/gluster/mnt8/brick
Brick25: s01-stg:/gluster/mnt9/brick
Brick26: s02-stg:/gluster/mnt9/brick
Brick27: s03-stg:/gluster/mnt9/brick
Brick28: s01-stg:/gluster/mnt10/brick
Brick29: s02-stg:/gluster/mnt10/brick
Brick30: s03-stg:/gluster/mnt10/brick
Brick31: s01-stg:/gluster/mnt11/brick
Brick32: s02-stg:/gluster/mnt11/brick
Brick33: s03-stg:/gluster/mnt11/brick
Brick34: s01-stg:/gluster/mnt12/brick
Brick35: s02-stg:/gluster/mnt12/brick
Brick36: s03-stg:/gluster/mnt12/brick
Options Reconfigured:
features.scrub: Active
features.bitrot: on
features.inode-quota: on
features.quota: on
performance.client-io-threads: on
cluster.min-free-disk: 10
cluster.quorum-type: auto
transport.address-family: inet
nfs.disable: on
server.event-threads: 4
client.event-threads: 4
cluster.lookup-optimize: on
performance.readdir-ahead: on
performance.parallel-readdir: off
cluster.readdir-optimize: on
features.cache-invalidation: on
features.cache-invalidation-timeout: 600
performance.stat-prefetch: on
performance.cache-invalidation: on
performance.md-cache-timeout: 600
network.inode-lru-limit: 50000
performance.io-cache: off
disperse.cpu-extensions: auto
performance.io-thread-count: 16
features.quota-deem-statfs: on
features.default-soft-limit: 90
cluster.server-quorum-type: server
cluster.brick-multiplex: on
cluster.server-quorum-ratio: 51%

[root@s01 ~]# rpm -qa|grep gluster
python2-gluster-3.10.5-1.el7.x86_64
glusterfs-geo-replication-3.10.5-1.el7.x86_64
centos-release-gluster310-1.0-1.el7.centos.noarch
glusterfs-server-3.10.5-1.el7.x86_64
glusterfs-libs-3.10.5-1.el7.x86_64
glusterfs-api-3.10.5-1.el7.x86_64
vdsm-gluster-4.19.31-1.el7.centos.noarch
glusterfs-3.10.5-1.el7.x86_64
gluster-nagios-common-1.1.0-0.el7.centos.noarch
glusterfs-cli-3.10.5-1.el7.x86_64
glusterfs-client-xlators-3.10.5-1.el7.x86_64
gluster-nagios-addons-1.1.0-0.el7.centos.x86_64
glusterfs-fuse-3.10.5-1.el7.x86_64
libvirt-daemon-driver-storage-gluster-3.2.0-14.el7_4.3.x86_64

Many thanks,
Mauro

> Il giorno 29 dic 2017, alle ore 04:59, Nithya Balachandran 
> <[email protected]> ha scritto:
> 
> Hi Mauro,
> 
> What version of Gluster are you running and what is your volume configuration?
> 
> IIRC, this was seen because of mismatches in the ctime returned to the 
> client. I don't think there were issues with the files but I will leave it to 
> Ravi and Raghavendra to comment.
> 
> 
> Regards,
> Nithya
> 
> 
> On 29 December 2017 at 04:10, Mauro Tridici <[email protected] 
> <mailto:[email protected]>> wrote:
> 
> Hi All,
> 
> anyone had the same experience?
> Could you provide me some information about this error?
> It happens only on GlusterFS file system.
> 
> Thank you,
> Mauro
> 
>> Il giorno 20 dic 2017, alle ore 16:57, Mauro Tridici <[email protected] 
>> <mailto:[email protected]>> ha scritto:
>> 
>> 
>> Dear Users,
>> 
>> I’m experiencing a random problem ( "file changed as we read it” error) 
>> during tar files creation on a distributed dispersed Gluster file system.
>> 
>> The tar files seem to be created correctly, but I can see a lot of message 
>> similar to the following ones:
>> 
>> tar: ./year1990/lffd1990050706p.nc <http://lffd1990050706p.nc/>.gz: file 
>> changed as we read it
>> tar: ./year1990/lffd1990052106p.nc <http://lffd1990052106p.nc/>.gz: file 
>> changed as we read it
>> tar: ./year1990/lffd1990052412p.nc <http://lffd1990052412p.nc/>.gz: file 
>> changed as we read it
>> tar: ./year1990/lffd1990091018.nc <http://lffd1990091018.nc/>.gz: file 
>> changed as we read it
>> tar: ./year1990/lffd1990092300p.nc <http://lffd1990092300p.nc/>.gz: file 
>> changed as we read it
>> tar: ./year1990/lffd1990092706p.nc <http://lffd1990092706p.nc/>.gz: file 
>> changed as we read it
>> tar: ./year1990/lffd1990100312p.nc <http://lffd1990100312p.nc/>.gz: file 
>> changed as we read it
>> tar: ./year1990/lffd1990100412.nc <http://lffd1990100412.nc/>.gz: file 
>> changed as we read it
>> tar: ./year1991/lffd1991012106.nc <http://lffd1991012106.nc/>.gz: file 
>> changed as we read it
>> tar: ./year1991/lffd1991010918.nc <http://lffd1991010918.nc/>.gz: file 
>> changed as we read it
>> tar: ./year1991/lffd1991011400.nc <http://lffd1991011400.nc/>.gz: file 
>> changed as we read it
>> 
>> I’m executing the tar command on a CentOS 6.2 operating system based server: 
>> it is a gluster native client.
>> 
>> You can find below some basic info about the gluster client:
>> 
>> [root@athena# rpm -qa|grep gluster
>> glusterfs-3.10.5-1.el6.x86_64
>> centos-release-gluster310-1.0-1.el6.centos.noarch
>> glusterfs-client-xlators-3.10.5-1.el6.x86_64
>> glusterfs-fuse-3.10.5-1.el6.x86_64
>> glusterfs-libs-3.10.5-1.el6.x86_64
>> 
>> Can I consider them as a false positive or the created tar files will suffer 
>> of inconsistence?
>> Is it a tar command problem or a gluster problem?
>> 
>> Could someone help me to resolve this issue?
>> 
>> Thank you very much,
>> Mauro
> 
> 
> 
> 
> 
> _______________________________________________
> Gluster-users mailing list
> [email protected] <mailto:[email protected]>
> http://lists.gluster.org/mailman/listinfo/gluster-users 
> <http://lists.gluster.org/mailman/listinfo/gluster-users>
> 


-------------------------
Mauro Tridici

Fondazione CMCC
CMCC Supercomputing Center
presso Complesso Ecotekne - Università del Salento -
Strada Prov.le Lecce - Monteroni sn
73100 Lecce  IT
http://www.cmcc.it

mobile: (+39) 327 5630841
email: [email protected]

_______________________________________________
Gluster-users mailing list
[email protected]
http://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to