In that case, most likely it seems to be an issue with the backup servers you are using.

Maybe you can first try verifying the NFS client on that machine. Issue write fops directly on the NFS mount points used by those servers.

Enable rpcdebug -> "rpcdebug -m nfs all" and check "/var/log/messages" for any errors.

Thanks,
Soumya

On 01/22/2015 11:11 PM, Peter Auyeung wrote:
Hi Soumya,

I was able to mount the same volume on other NFS client and do writes

got the following nfs.log entries when write




[2015-01-22 17:39:03.528405] I 
[afr-self-heal-common.c:2868:afr_log_self_heal_completion_status] 
0-sas02-replicate-1:  metadata self heal  is successfully completed,   metadata 
self heal from source sas02-client-2 to sas02-client-3,  metadata - Pending 
matrix:  [ [ 0 0 ] [ 0 0 ] ], on /RepDBSata02
[2015-01-22 17:39:03.529407] I 
[afr-self-heal-common.c:2868:afr_log_self_heal_completion_status] 
0-sas02-replicate-2:  metadata self heal  is successfully completed,   metadata 
self heal from source sas02-client-4 to sas02-client-5,  metadata - Pending 
matrix:  [ [ 0 0 ] [ 0 0 ] ], on /RepDBSata02


Thanks
Peter
________________________________________
From: Soumya Koduri [[email protected]]
Sent: Wednesday, January 21, 2015 9:05 PM
To: Peter Auyeung; [email protected]; [email protected]
Subject: Re: [Gluster-devel] [Gluster-user] Sybase backup server failed to 
write to Gluster NFS

Hi Peter,

Can you please try manually mounting those volumes using any/other nfs
client and check if you are able to perform write operations. Also
please collect the gluster nfs log while doing so.

Thanks,
Soumya

On 01/22/2015 08:18 AM, Peter Auyeung wrote:
Hi,

We have been having 5 sybase servers doing dump/export to Gluster NFS
for couple months and yesterday it started to give us these error on not
able to write files

The gluster NFS export is not full and we can still move and write files
as sybase unix user from the sybase servers.

There are no error logs on gluster nfs nor the bricks and etc-glusterfs
logs and no nfs client error on the sybase servers neither.

The NFS export was a replica 2 volume (3x2)

I created another NFS export from same gluster but a distributed only
volume and still giving out the same error.

Any Clue?

Thanks
Peter

Jan 20 20:04:17 2015: Backup Server: 6.53.1.1: OPERATOR: Volume on
device '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.e'
cannot be opened for write access. Mount another volume.
Jan 20 20:04:17 2015: Backup Server: 6.78.1.1: EXECUTE sp_volchanged
          @session_id = 87,
          @devname =
'/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.e',
          @action = { 'PROCEED' | 'RETRY' | 'ABORT' }
Jan 20 20:04:26 2015: Backup Server: 6.53.1.1: OPERATOR: Volume on
device '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.a'
cannot be opened for write access. Mount another volume.
Jan 20 20:04:26 2015: Backup Server: 6.78.1.1: EXECUTE sp_volchanged
          @session_id = 87,
          @devname =
'/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.a',
          @action = { 'PROCEED' | 'RETRY' | 'ABORT' }
Jan 20 20:05:41 2015: Backup Server: 6.53.1.1: OPERATOR: Volume on
device '/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.d'
cannot be opened for write access. Mount another volume.
Jan 20 20:05:41 2015: Backup Server: 6.78.1.1: EXECUTE sp_volchanged
          @session_id = 87,
          @devname =
'/dbbackup01/db/full/pr_rssd_id_repsrv_rssd.F01-20-20-04.d',
          @action = { 'PROCEED' | 'RETRY' | 'ABORT' }



_______________________________________________
Gluster-devel mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-devel

_______________________________________________
Gluster-devel mailing list
[email protected]
http://www.gluster.org/mailman/listinfo/gluster-devel

Reply via email to