On 09/07/13 18:17, 符永涛 wrote:
Hi Toby,

What's the bug #? I want to have a look and backport it to our
production server if it helps. Thank you.

I think it was this one:
https://bugzilla.redhat.com/show_bug.cgi?id=947824

The bug being that the daemons were crashing out if you had a lot of volumes defined, I think?

Toby

2013/7/9 Toby Corkindale <[email protected]
<mailto:[email protected]>>

    On 09/07/13 15:38, Bobby Jacob wrote:

        Hi,

        I have a 2-node gluster with 3 TB storage.

        1)I believe the “glusterfsd” is responsible for the self healing
        between
        the 2 nodes.

        2)Due to some network error, the replication stopped for some
        reason but

        the application was accessing the data from node1.  When I
        manually try
        to start “glusterfsd” service, its not starting.

        Please advice on how I can maintain the integrity of the data so
        that we
        have all the data in both the locations. ??


    There were some bugs in the self-heal daemon present in 3.3.0 and
    3.3.1. Our systems see the SHD crash out with segfaults quite often,
    and it does not recover.

    I reported this bug a long time ago, and it was fixed in trunk
    relatively quickly -- however version 3.3.2 has still not been
    released, despite the fix being found six months ago.

    I find this quite disappointing.

    T
    _________________________________________________
    Gluster-users mailing list
    [email protected] <mailto:[email protected]>
    http://supercolony.gluster.__org/mailman/listinfo/gluster-__users
    <http://supercolony.gluster.org/mailman/listinfo/gluster-users>




--
符永涛

_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Reply via email to