On 02/06/2015 08:11 AM, Humble Devassy Chirammal wrote:
On 02/05/2015 11:56 PM, Nux! wrote:
Thanks for sharing.
Any idea if 3.6.2 still is compatible with v3.4 servers?
>You mean 3.6.2 client bits with v3.4 servers? yes, it should work fine.
afacit, this will *not* work and its *not* supported.
Humble,
Gluster 3.4 server bits with 3.6 client bits should work fine.
But I think the reserve (i.e. 3.6 server bits with older client bits)
are not compatible because of below issues
* Older clients can not mount the newly created volume on 3.6 . This
is because readdir-ahead will be enabled on the volume by default
which isn't present in older clients.
* We can't run rebalance on any volume created with 3.6 bits ( with
or without readdir-ahead) when older clients are connected. The
rebalance command will error out if older clients are connected.
Thanks,
Lala
--Humble
On Fri, Feb 6, 2015 at 5:02 AM, Lalatendu Mohanty <lmoha...@redhat.com
<mailto:lmoha...@redhat.com>> wrote:
+ gluster-users
On 02/05/2015 11:56 PM, Nux! wrote:
Thanks for sharing.
Any idea if 3.6.2 still is compatible with v3.4 servers?
You mean 3.6.2 client bits with v3.4 servers? yes, it should work
fine.
-Lala
--
Sent from the Delta quadrant using Borg technology!
Nux!
www.nux.ro <http://www.nux.ro>
----- Original Message -----
From: "Karanbir Singh" <mail-li...@karan.org
<mailto:mail-li...@karan.org>>
To: "The CentOS developers mailing list."
<centos-de...@centos.org <mailto:centos-de...@centos.org>>
Sent: Thursday, 5 February, 2015 22:11:53
Subject: [CentOS-devel] Gluster Updates for Storage SIG
The CentOS Storage SIG, has updated Gluster to 3.6.2 in
the community
testing repos. You can find more information on howto get
started with
this repo at :
http://wiki.centos.org/SpecialInterestGroup/Storage/gluster-Quickstart
The Following rpms have been updated:
CentOS-6
i386/glusterfs-3.6.2-2.el6.i386.rpm
i386/glusterfs-api-3.6.2-2.el6.i386.rpm
i386/glusterfs-api-devel-3.6.2-2.el6.i386.rpm
i386/glusterfs-cli-3.6.2-2.el6.i386.rpm
i386/glusterfs-devel-3.6.2-2.el6.i386.rpm
i386/glusterfs-extra-xlators-3.6.2-2.el6.i386.rpm
i386/glusterfs-fuse-3.6.2-2.el6.i386.rpm
i386/glusterfs-geo-replication-3.6.2-2.el6.i386.rpm
i386/glusterfs-libs-3.6.2-2.el6.i386.rpm
i386/glusterfs-rdma-3.6.2-2.el6.i386.rpm
i386/glusterfs-resource-agents-3.6.2-2.el6.noarch.rpm
i386/glusterfs-server-3.6.2-2.el6.i386.rpm
x86_64/glusterfs-3.6.2-2.el6.x86_64.rpm
x86_64/glusterfs-api-3.6.2-2.el6.i386.rpm
x86_64/glusterfs-api-3.6.2-2.el6.x86_64.rpm
x86_64/glusterfs-api-devel-3.6.2-2.el6.i386.rpm
x86_64/glusterfs-api-devel-3.6.2-2.el6.x86_64.rpm
x86_64/glusterfs-cli-3.6.2-2.el6.x86_64.rpm
x86_64/glusterfs-devel-3.6.2-2.el6.i386.rpm
x86_64/glusterfs-devel-3.6.2-2.el6.x86_64.rpm
x86_64/glusterfs-extra-xlators-3.6.2-2.el6.x86_64.rpm
x86_64/glusterfs-fuse-3.6.2-2.el6.x86_64.rpm
x86_64/glusterfs-geo-replication-3.6.2-2.el6.x86_64.rpm
x86_64/glusterfs-libs-3.6.2-2.el6.i386.rpm
x86_64/glusterfs-libs-3.6.2-2.el6.x86_64.rpm
x86_64/glusterfs-rdma-3.6.2-2.el6.x86_64.rpm
x86_64/glusterfs-resource-agents-3.6.2-2.el6.noarch.rpm
x86_64/glusterfs-server-3.6.2-2.el6.x86_64.rpm
CentOS-7
x86_64/glusterfs-3.6.2-2.el7.x86_64.rpm
x86_64/glusterfs-api-3.6.2-2.el7.x86_64.rpm
x86_64/glusterfs-api-devel-3.6.2-2.el7.x86_64.rpm
x86_64/glusterfs-cli-3.6.2-2.el7.x86_64.rpm
x86_64/glusterfs-devel-3.6.2-2.el7.x86_64.rpm
x86_64/glusterfs-extra-xlators-3.6.2-2.el7.x86_64.rpm
x86_64/glusterfs-fuse-3.6.2-2.el7.x86_64.rpm
x86_64/glusterfs-geo-replication-3.6.2-2.el7.x86_64.rpm
x86_64/glusterfs-libs-3.6.2-2.el7.x86_64.rpm
x86_64/glusterfs-rdma-3.6.2-2.el7.x86_64.rpm
x86_64/glusterfs-resource-agents-3.6.2-2.el7.noarch.rpm
x86_64/glusterfs-server-3.6.2-2.el7.x86_64.rpm
This release fixes the following bugs. Below containt
copied from
GlusterFS upstream release mail [1].
1184191 - Cluster/DHT : Fixed crash due to null deref
1180404 - nfs server restarts when a snapshot is deactivated
1180411 - CIFS:[USS]: glusterfsd OOM killed when 255
snapshots were
browsed at CIFS mount and Control+C is issued
1180070 - [AFR] getfattr on fuse mount gives error :
Software caused
connection abort
1175753 - [readdir-ahead]: indicate EOF for readdirp
1175752 - [USS]: On a successful lookup, snapd logs are
filled with
Warnings "dict OR key (entry-point) is NULL"
1175749 - glusterfs client crashed while migrating the fds
1179658 - Add brick fails if parent dir of new brick and
existing brick
is same and volume was accessed using libgfapi and smb.
1146524 - glusterfs.spec.in <http://glusterfs.spec.in> -
synch minor diffs with fedora dist-git
glusterfs.spec
1175744 - [USS]: Unable to access .snaps after snapshot
restore after
directories were deleted and recreated
1175742 - [USS]: browsing .snaps directory with CIFS fails
with
"Invalid argument"
1175739 - [USS]: Non root user who has no access to a
directory, from
NFS mount, is able to access the files under .snaps under
that directory
1175758 - [USS] : Rebalance process tries to connect to
snapd and in
case when snapd crashes it might affect rebalance process
1175765 - USS]: When snapd is crashed gluster volume
stop/delete
operation fails making the cluster in inconsistent state
1173528 - Change in volume heal info command output
1166515 - [Tracker] RDMA support in glusterfs
1166505 - mount fails for nfs protocol in rdma volumes
1138385 - [DHT:REBALANCE]: Rebalance failures are seen
with error
message " remote operation failed: File exists"
1177418 - entry self-heal in 3.5 and 3.6 are not compatible
1170954 - Fix mutex problems reported by coverity scan
1177899 - nfs: ls shows "Permission denied" with root-squash
1175738 - [USS]: data unavailability for a period of time
when USS is
enabled/disabled
1175736 - [USS]:After deactivating a snapshot trying to
access the
remaining activated snapshots from NFS mount gives
'Invalid argument' error
1175735 - [USS]: snapd process is not killed once the
glusterd comes back
1175733 - [USS]: If the snap name is same as
snap-directory than cd to
virtual snap directory fails
1175756 - [USS] : Snapd crashed while trying to access the
snapshots
under .snaps directory
1175755 - SNAPSHOT[USS]:gluster volume set for uss doesnot
check any
boundaries
1175732 - [SNAPSHOT]: nouuid is appended for every
snapshoted brick
which causes duplication if the original brick has already
nouuid
1175730 - [USS]: creating file/directories under .snaps
shows wrong
error message
1175754 - [SNAPSHOT]: before the snap is marked to be
deleted if the
node goes down than the snaps are propagated on other
nodes and glusterd
hungs
1159484 - ls -alR can not heal the disperse volume
1138897 - NetBSD port
1175728 - [USS]: All uss related logs are reported under
/var/log/glusterfs, it makes sense to move it into subfolder
1170548 - [USS] : don't display the snapshots which are
not activated
1170921 - [SNAPSHOT]: snapshot should be deactivated by
default when
created
1175694 - [SNAPSHOT]: snapshoted volume is read only but
it shows rw
attributes in mount
1161885 - Possible file corruption on dispersed volumes
1170959 - EC_MAX_NODES is defined incorrectly
1175645 - [USS]: Typo error in the description for USS
under "gluster
volume set help"
1171259 - mount.glusterfs does not understand -n option
[1]
http://www.gluster.org/pipermail/gluster-devel/2015-January/043617.html
--
Karanbir Singh
+44-207-0999389 | http://www.karan.org/ |
twitter.com/kbsingh <http://twitter.com/kbsingh>
GnuPG Key : http://www.karan.org/publickey.asc
_______________________________________________
CentOS-devel mailing list
centos-de...@centos.org <mailto:centos-de...@centos.org>
http://lists.centos.org/mailman/listinfo/centos-devel
_______________________________________________
CentOS-devel mailing list
centos-de...@centos.org <mailto:centos-de...@centos.org>
http://lists.centos.org/mailman/listinfo/centos-devel
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org <mailto:Gluster-users@gluster.org>
http://www.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users