Lala,
Gluster 3.4 server bits with 3.6 client bits should work fine.
Have you tested this configuration? iic, the 'AFR' module ( replication) introduced its 'Version 2' implementation in 3.6 which is not compatible with its older version. The GlusterFS 3.4 & 3.5 versions are shipped with "AFR V1" , so I really doubt the mentioned configuration will work perfectly. AFR guys can confirm though.
--Humble
On Fri, Feb 6, 2015 at 7:01 PM, Lalatendu Mohanty lmohanty@redhat.com wrote:
On 02/06/2015 08:11 AM, Humble Devassy Chirammal wrote:
On 02/05/2015 11:56 PM, Nux! wrote:
Thanks for sharing. Any idea if 3.6.2 still is compatible with v3.4 servers?
You mean 3.6.2 client bits with v3.4 servers? yes, it should work fine.
afacit, this will *not* work and its *not* supported.
Humble,
Gluster 3.4 server bits with 3.6 client bits should work fine.
But I think the reserve (i.e. 3.6 server bits with older client bits) are not compatible because of below issues
- Older clients can not mount the newly created volume on 3.6 . This
is because readdir-ahead will be enabled on the volume by default which isn't present in older clients. - We can't run rebalance on any volume created with 3.6 bits ( with or without readdir-ahead) when older clients are connected. The rebalance command will error out if older clients are connected.
Thanks, Lala
--Humble
On Fri, Feb 6, 2015 at 5:02 AM, Lalatendu Mohanty lmohanty@redhat.com wrote:
- gluster-users
On 02/05/2015 11:56 PM, Nux! wrote:
Thanks for sharing. Any idea if 3.6.2 still is compatible with v3.4 servers?
You mean 3.6.2 client bits with v3.4 servers? yes, it should work fine.
-Lala
--
Sent from the Delta quadrant using Borg technology!
Nux! www.nux.ro
----- Original Message -----
From: "Karanbir Singh" mail-lists@karan.org To: "The CentOS developers mailing list." centos-devel@centos.org Sent: Thursday, 5 February, 2015 22:11:53 Subject: [CentOS-devel] Gluster Updates for Storage SIG The CentOS Storage SIG, has updated Gluster to 3.6.2 in the community testing repos. You can find more information on howto get started with this repo at : http://wiki.centos.org/SpecialInterestGroup/Storage/gluster-Quickstart
The Following rpms have been updated:
CentOS-6 i386/glusterfs-3.6.2-2.el6.i386.rpm i386/glusterfs-api-3.6.2-2.el6.i386.rpm i386/glusterfs-api-devel-3.6.2-2.el6.i386.rpm i386/glusterfs-cli-3.6.2-2.el6.i386.rpm i386/glusterfs-devel-3.6.2-2.el6.i386.rpm i386/glusterfs-extra-xlators-3.6.2-2.el6.i386.rpm i386/glusterfs-fuse-3.6.2-2.el6.i386.rpm i386/glusterfs-geo-replication-3.6.2-2.el6.i386.rpm i386/glusterfs-libs-3.6.2-2.el6.i386.rpm i386/glusterfs-rdma-3.6.2-2.el6.i386.rpm i386/glusterfs-resource-agents-3.6.2-2.el6.noarch.rpm i386/glusterfs-server-3.6.2-2.el6.i386.rpm
x86_64/glusterfs-3.6.2-2.el6.x86_64.rpm x86_64/glusterfs-api-3.6.2-2.el6.i386.rpm x86_64/glusterfs-api-3.6.2-2.el6.x86_64.rpm x86_64/glusterfs-api-devel-3.6.2-2.el6.i386.rpm x86_64/glusterfs-api-devel-3.6.2-2.el6.x86_64.rpm x86_64/glusterfs-cli-3.6.2-2.el6.x86_64.rpm x86_64/glusterfs-devel-3.6.2-2.el6.i386.rpm x86_64/glusterfs-devel-3.6.2-2.el6.x86_64.rpm x86_64/glusterfs-extra-xlators-3.6.2-2.el6.x86_64.rpm x86_64/glusterfs-fuse-3.6.2-2.el6.x86_64.rpm x86_64/glusterfs-geo-replication-3.6.2-2.el6.x86_64.rpm x86_64/glusterfs-libs-3.6.2-2.el6.i386.rpm x86_64/glusterfs-libs-3.6.2-2.el6.x86_64.rpm x86_64/glusterfs-rdma-3.6.2-2.el6.x86_64.rpm x86_64/glusterfs-resource-agents-3.6.2-2.el6.noarch.rpm x86_64/glusterfs-server-3.6.2-2.el6.x86_64.rpm
CentOS-7 x86_64/glusterfs-3.6.2-2.el7.x86_64.rpm x86_64/glusterfs-api-3.6.2-2.el7.x86_64.rpm x86_64/glusterfs-api-devel-3.6.2-2.el7.x86_64.rpm x86_64/glusterfs-cli-3.6.2-2.el7.x86_64.rpm x86_64/glusterfs-devel-3.6.2-2.el7.x86_64.rpm x86_64/glusterfs-extra-xlators-3.6.2-2.el7.x86_64.rpm x86_64/glusterfs-fuse-3.6.2-2.el7.x86_64.rpm x86_64/glusterfs-geo-replication-3.6.2-2.el7.x86_64.rpm x86_64/glusterfs-libs-3.6.2-2.el7.x86_64.rpm x86_64/glusterfs-rdma-3.6.2-2.el7.x86_64.rpm x86_64/glusterfs-resource-agents-3.6.2-2.el7.noarch.rpm x86_64/glusterfs-server-3.6.2-2.el7.x86_64.rpm
This release fixes the following bugs. Below containt copied from GlusterFS upstream release mail [1].
1184191 - Cluster/DHT : Fixed crash due to null deref 1180404 - nfs server restarts when a snapshot is deactivated 1180411 - CIFS:[USS]: glusterfsd OOM killed when 255 snapshots were browsed at CIFS mount and Control+C is issued 1180070 - [AFR] getfattr on fuse mount gives error : Software caused connection abort 1175753 - [readdir-ahead]: indicate EOF for readdirp 1175752 - [USS]: On a successful lookup, snapd logs are filled with Warnings "dict OR key (entry-point) is NULL" 1175749 - glusterfs client crashed while migrating the fds 1179658 - Add brick fails if parent dir of new brick and existing brick is same and volume was accessed using libgfapi and smb. 1146524 - glusterfs.spec.in - synch minor diffs with fedora dist-git glusterfs.spec 1175744 - [USS]: Unable to access .snaps after snapshot restore after directories were deleted and recreated 1175742 - [USS]: browsing .snaps directory with CIFS fails with "Invalid argument" 1175739 - [USS]: Non root user who has no access to a directory, from NFS mount, is able to access the files under .snaps under that directory 1175758 - [USS] : Rebalance process tries to connect to snapd and in case when snapd crashes it might affect rebalance process 1175765 - USS]: When snapd is crashed gluster volume stop/delete operation fails making the cluster in inconsistent state 1173528 - Change in volume heal info command output 1166515 - [Tracker] RDMA support in glusterfs 1166505 - mount fails for nfs protocol in rdma volumes 1138385 - [DHT:REBALANCE]: Rebalance failures are seen with error message " remote operation failed: File exists" 1177418 - entry self-heal in 3.5 and 3.6 are not compatible 1170954 - Fix mutex problems reported by coverity scan 1177899 - nfs: ls shows "Permission denied" with root-squash 1175738 - [USS]: data unavailability for a period of time when USS is enabled/disabled 1175736 - [USS]:After deactivating a snapshot trying to access the remaining activated snapshots from NFS mount gives 'Invalid argument' error 1175735 - [USS]: snapd process is not killed once the glusterd comes back 1175733 - [USS]: If the snap name is same as snap-directory than cd to virtual snap directory fails 1175756 - [USS] : Snapd crashed while trying to access the snapshots under .snaps directory 1175755 - SNAPSHOT[USS]:gluster volume set for uss doesnot check any boundaries 1175732 - [SNAPSHOT]: nouuid is appended for every snapshoted brick which causes duplication if the original brick has already nouuid 1175730 - [USS]: creating file/directories under .snaps shows wrong error message 1175754 - [SNAPSHOT]: before the snap is marked to be deleted if the node goes down than the snaps are propagated on other nodes and glusterd hungs 1159484 - ls -alR can not heal the disperse volume 1138897 - NetBSD port 1175728 - [USS]: All uss related logs are reported under /var/log/glusterfs, it makes sense to move it into subfolder 1170548 - [USS] : don't display the snapshots which are not activated 1170921 - [SNAPSHOT]: snapshot should be deactivated by default when created 1175694 - [SNAPSHOT]: snapshoted volume is read only but it shows rw attributes in mount 1161885 - Possible file corruption on dispersed volumes 1170959 - EC_MAX_NODES is defined incorrectly 1175645 - [USS]: Typo error in the description for USS under "gluster volume set help" 1171259 - mount.glusterfs does not understand -n option
[1] http://www.gluster.org/pipermail/gluster-devel/2015-January/043617.html
-- Karanbir Singh +44-207-0999389 | http://www.karan.org/ | twitter.com/kbsingh GnuPG Key : http://www.karan.org/publickey.asc _______________________________________________ CentOS-devel mailing list CentOS-devel@centos.org http://lists.centos.org/mailman/listinfo/centos-devel
CentOS-devel mailing list CentOS-devel@centos.org http://lists.centos.org/mailman/listinfo/centos-devel
Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users