On 09/07/13 18:17, 符永涛 wrote:
Hi Toby,
What's the bug #? I want to have a look and backport it to our
production server if it helps. Thank you.
I think it was this one:
https://bugzilla.redhat.com/show_bug.cgi?id=947824
The bug being that the daemons were crashing out if you had a lot of
volumes defined, I think?
Toby
2013/7/9 Toby Corkindale <[email protected]
<mailto:[email protected]>>
On 09/07/13 15:38, Bobby Jacob wrote:
Hi,
I have a 2-node gluster with 3 TB storage.
1)I believe the “glusterfsd” is responsible for the self healing
between
the 2 nodes.
2)Due to some network error, the replication stopped for some
reason but
the application was accessing the data from node1. When I
manually try
to start “glusterfsd” service, its not starting.
Please advice on how I can maintain the integrity of the data so
that we
have all the data in both the locations. ??
There were some bugs in the self-heal daemon present in 3.3.0 and
3.3.1. Our systems see the SHD crash out with segfaults quite often,
and it does not recover.
I reported this bug a long time ago, and it was fixed in trunk
relatively quickly -- however version 3.3.2 has still not been
released, despite the fix being found six months ago.
I find this quite disappointing.
T
_________________________________________________
Gluster-users mailing list
[email protected] <mailto:[email protected]>
http://supercolony.gluster.__org/mailman/listinfo/gluster-__users
<http://supercolony.gluster.org/mailman/listinfo/gluster-users>
--
符永涛
_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users